blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 3 281 | content_id stringlengths 40 40 | detected_licenses listlengths 0 57 | license_type stringclasses 2 values | repo_name stringlengths 6 116 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringclasses 313 values | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 18.2k 668M ⌀ | star_events_count int64 0 102k | fork_events_count int64 0 38.2k | gha_license_id stringclasses 17 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 107 values | src_encoding stringclasses 20 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 4 6.02M | extension stringclasses 78 values | content stringlengths 2 6.02M | authors listlengths 1 1 | author stringlengths 0 175 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
714db4e6784f72403c33549680bae507b0d5bb17 | b951b28a8957cc788ea297a294766117b6ee9e04 | /task_2.py | a6aef9d9e9ef01ef8c7cbd86d178b29c6a4962f5 | [] | no_license | goodprogrammist/module3 | ada4b757a1020614b86a6e36eca9718aa6aeb546 | 4dd60fbf7b948dfa4398406960a73a25d3eb6ec6 | refs/heads/master | 2023-04-20T22:37:58.072734 | 2021-05-15T15:26:08 | 2021-05-15T15:26:08 | 367,620,899 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 959 | py | print('Задача 2. Финансовый отчёт')
# Наде дали задание сформировать финансовый отчёт за последние 20 лет по полугодиям.
# Нужно сумму дохода первых двух кварталов поделить на сумму последних двух кварталов,
# чтобы понять динамику роста или падения дохода. И так за каждый год.
#
# Надя решила,
# что быстрее будет написать простую программу, которая сделает всё за неё.
#
#
# Запросите у пользователя четыре числа.
# Отдельно сложите первые два и отдельно вторые два.
# Разделите первую сумму на вторую.
# Выведите результат на экран.
| [
""
] | |
50eb91f0fbd18a59b7eb5e494dc487cac5caadee | 821985c62bf469b72a7d451560a455aefe2c7bd0 | /A、机器学习/机器学习实践/RandomForest/4.RandomForestClassifier_Enum_iris.py | bc292c16243930d4ddca12758d85363a76bc9f66 | [] | no_license | FangChao1086/Machine_Learning | 625106c977055328368100dae30f00909aff590c | 603900b510a9a64f4bb6fcc7753546b34828e022 | refs/heads/master | 2020-04-16T13:20:53.675722 | 2020-03-14T03:33:21 | 2020-03-14T03:33:21 | 165,622,654 | 14 | 2 | null | null | null | null | UTF-8 | Python | false | false | 2,770 | py | #!/usr/bin/python
# -*- coding:utf-8 -*-
import numpy as np
import matplotlib.pyplot as plt
import matplotlib as mpl
from sklearn.ensemble import RandomForestClassifier
def iris_type(s):
it = {b'Iris-setosa': 0, b'Iris-versicolor': 1, b'Iris-virginica': 2}
return it[s]
# 'sepal length', 'sepal width', 'petal length', 'petal width'
iris_feature = u'花萼长度', u'花萼宽度', u'花瓣长度', u'花瓣宽度'
if __name__ == "__main__":
mpl.rcParams['font.sans-serif'] = [u'SimHei'] # 黑体 FangSong/KaiTi
mpl.rcParams['axes.unicode_minus'] = False
path = 'DATASET_iris.data' # 数据文件路径
data = np.loadtxt(path, dtype=float, delimiter=',', converters={4: iris_type})
x_prime, y = np.split(data, (4,), axis=1)
feature_pairs = [[0, 1], [0, 2], [0, 3], [1, 2], [1, 3], [2, 3]]
plt.figure(figsize=(10, 9), facecolor='#FFFFFF')
for i, pair in enumerate(feature_pairs):
# 准备数据
x = x_prime[:, pair]
# 随机森林
clf = RandomForestClassifier(n_estimators=200, criterion='entropy', max_depth=3)
rf_clf = clf.fit(x, y.ravel())
# 画图
N, M = 50, 50 # 横纵各采样多少个值
x1_min, x1_max = x[:, 0].min(), x[:, 0].max() # 第0列的范围
x2_min, x2_max = x[:, 1].min(), x[:, 1].max() # 第1列的范围
t1 = np.linspace(x1_min, x1_max, N)
t2 = np.linspace(x2_min, x2_max, M)
x1, x2 = np.meshgrid(t1, t2) # 生成网格采样点
x_test = np.stack((x1.flat, x2.flat), axis=1) # 测试点
# 训练集上的预测结果
y_hat = rf_clf.predict(x)
y = y.reshape(-1)
c = np.count_nonzero(y_hat == y) # 统计预测正确的个数
print('特征: ', iris_feature[pair[0]], ' + ', iris_feature[pair[1]],)
print('预测正确数目:', c,)
print('准确率: %.2f%%' % (100 * float(c) / float(len(y))))
# 显示
cm_light = mpl.colors.ListedColormap(['#A0FFA0', '#FFA0A0', '#A0A0FF'])
cm_dark = mpl.colors.ListedColormap(['g', 'r', 'b'])
y_hat = rf_clf.predict(x_test) # 预测值
y_hat = y_hat.reshape(x1.shape) # 使之与输入的形状相同
plt.subplot(2, 3, i+1)
plt.pcolormesh(x1, x2, y_hat, cmap=cm_light) # 预测值
plt.scatter(x[:, 0], x[:, 1], c=y, edgecolors='k', cmap=cm_dark) # 样本
plt.xlabel(iris_feature[pair[0]], fontsize=14)
plt.ylabel(iris_feature[pair[1]], fontsize=14)
plt.xlim(x1_min, x1_max)
plt.ylim(x2_min, x2_max)
plt.grid()
plt.tight_layout(2.5)
plt.subplots_adjust(top=0.92)
plt.suptitle(u'随机森林对鸢尾花数据的两特征组合的分类结果', fontsize=18)
plt.show()
| [
"“fangchao1086@163.com"
] | “fangchao1086@163.com |
cc9a5cfca7fa31d2f2c94d92b36a1c1bd5caae4b | a9e6a1dea27a984b7e3f46c56e5d62e57ac5d8a5 | /first/gevent/3.py | 8557daca47a4c80492bf9f028e91463ad466bcb9 | [] | no_license | spiderTom/pythonReference | fd2f21c0c3b919e08429b27464b4d161a20db2f7 | fa1d7de8630c12a8add3a7370a452daef0e62c73 | refs/heads/master | 2021-01-19T05:01:01.490392 | 2017-04-06T09:01:04 | 2017-04-06T09:01:04 | 87,408,456 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 129 | py |
def fibon(n):
a = b = 1
for i in range(n):
yield a
a, b = b, a + b
for x in fibon(20):
print(x)
| [
"tomtubo@163.com"
] | tomtubo@163.com |
987811a6ffc6d66edfc20e22d3a6be53f16d4372 | 0b6dade21d327773ab958db6b85af3ef68dfbbe0 | /hello.py | bad42797591e2329f10c59a41eb9f9c0fd568a41 | [
"Apache-2.0"
] | permissive | maharsh-cmput404/cgi-lab | 8e10ba1762dee5650d2911c032b94e1dd3decd46 | b8b5a2bf15b7e689d7507f89e57506ab718e6863 | refs/heads/master | 2020-12-15T19:20:40.657703 | 2020-01-21T00:49:43 | 2020-01-21T00:49:43 | 235,226,103 | 0 | 0 | null | 2020-01-21T00:39:14 | 2020-01-21T00:39:13 | null | UTF-8 | Python | false | false | 202 | py | #!/usr/bin/env python3
import os
import cgi
import cgitb
cgitb.enable()
print("Content-Type: text/plain\n")
print()
# print("<!doctype html><title>Hello</title><h2>Hello World</h2>")
print(os.environ)
| [
"patel@maharsh.net"
] | patel@maharsh.net |
4b2f6eff5f5301a631de8ff075d91eec3458cd2e | 7f43804828706cfd3635c1f1d8cfff1c7fc33beb | /2019/day08/day8.py | c42c2ca2f84644ceb98f510cc13bac1a3db0501b | [] | no_license | aldewereld/AventOfCode | b9149a0161d89cd550c53b20a3a42f069b520e41 | 2da755700593e98445736d23b16dfb10483df049 | refs/heads/master | 2020-09-23T08:44:25.553145 | 2019-12-12T13:43:14 | 2019-12-12T13:43:14 | 225,455,849 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,527 | py | from typing import List, Tuple
from functools import reduce
with open("input.txt") as file:
line = list(file.readline().strip())
def split_layers(input_message: List[str], width: int = 25, height: int = 6) -> List[List[str]]:
layers = []
start = 0
step = width * height
end = step
for i in range(0, len(input_message), step):
layers.append(input_message[start:end])
start, end = end, end+step
return layers
def count(object: int, list: List[int]) -> int:
if list == []:
return 0
else:
head, *tail = list
if head == object:
return 1+count(object, tail)
else:
return count(object, tail)
# Part 1
layers = split_layers(line)
counted = list(map(lambda x: count('0', x), layers))
least_zeros = min(counted)
index = counted.index(least_zeros)
ones, twos = count('1', layers[index]), count('2', layers[index])
print(ones * twos)
# Part 2
#layers = split_layers(list("0222112222120000"), 2, 2)
def decode_layers(layer1: List[int], layer2: List[int]) -> List[int]:
return list(map(decode_pixel, zip(layer1, layer2)))
def decode_pixel(pixel_pair: Tuple[int, int]) -> int:
if pixel_pair[0] == '2': # if transparent
return pixel_pair[1] # return lower pixel
else: # else use upper pixel
return pixel_pair[0]
image = reduce(decode_layers, layers)
start = 0
end = 25
for _ in range(6):
print(image[start:end])
start, end = end, end+25 | [
"huib.aldewereld@hu.nl"
] | huib.aldewereld@hu.nl |
9dea79ebe2acef41d229a77657e6b1cf232caf43 | 5215715a4cbcf9ce065b1542db224a2b1997c760 | /T3/t3.py | 5523731288ed9eb2d6e7a4f542b5a35d71a18b89 | [
"LicenseRef-scancode-warranty-disclaimer",
"MIT"
] | permissive | Cgipson06/reddit | 02deac29ead779890e42d48400d2233ce888e5a0 | deb1da398840bbd311a79eec25ef2a8b5a8ed5b1 | refs/heads/master | 2021-01-08T23:19:55.245559 | 2014-12-28T20:52:26 | 2014-12-28T20:52:26 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,034 | py | #/u/GoldenSights
import praw
import time
import sqlite3
import datetime
import random
USERAGENT = """
/u/GoldenSights T3 data collection: Gathering Submission data for
statistical analysis.
More info at https://github.com/voussoir/reddit/tree/master/T3
"""
r = praw.Reddit(USERAGENT)
print('Connected to reddit.')
sql = sqlite3.connect('D:/T3/t3.db')
cur = sql.cursor()
cur.execute('CREATE TABLE IF NOT EXISTS meta(label TEXT, data TEXT)')
cur.execute(('CREATE TABLE IF NOT EXISTS posts(idint INT, idstr TEXT, '
'created INT, self INT, nsfw INT, author TEXT, title TEXT, '
'url TEXT, selftext TEXT, score INT, subreddit TEXT, distinguish INT, '
'textlen INT)'))
DISTINGUISHMAP = {0:"user", 1:"moderator", 2:"admin"}
DISTINGUISHMAP_R = {"user":0, "moderator":1, "admin":2}
LOWERBOUND = 9999000
# 5yba0
UPPERBOUND = 164790958
# 2q41im
# 1,679,616 = 10000
# 9,999,000 = 5yba0
# 60,466,176 = 100000
# 120,932,352 = 200000
# 164,790,958 = 2q41im
# 0 - idint
# 1 - idstr
# 2 - created
# 3 - self
# 4 - nsfw
# 5 - author
# 6 - title
# 7 - url
# 8 - selftext
# 9 - score
# 10 - subreddit
# 11 - distinguished
# 12 - textlen
class Post:
''' Used to map the indices of DB entries to names '''
def __init__(self, data):
self.idint = data[0]
self.idstr = data[1]
self.created_utc = data[2]
self.is_self = True if data[3] == 1 else False
self.over_18 = True if data[4] == 1 else False
self.author = data[5]
self.title = data[6]
self.url = data[7]
self.selftext = data[8]
self.score = data[9]
self.subreddit = data[10]
self.distinguished = DISTINGUISHMAP[data[11]]
self.textlen = data[12]
def base36encode(number, alphabet='0123456789abcdefghijklmnopqrstuvwxyz'):
"""Converts an integer to a base36 string."""
if not isinstance(number, (int)):
raise TypeError('number must be an integer')
base36 = ''
sign = ''
if number < 0:
sign = '-'
number = -number
if 0 <= number < len(alphabet):
return sign + alphabet[number]
while number != 0:
number, i = divmod(number, len(alphabet))
base36 = alphabet[i] + base36
return sign + base36
def base36decode(number):
return int(number, 36)
def b36(i):
if type(i) == int:
return base36encode(i)
if type(i) == str:
return base36decode(i)
def human(timestamp):
day = datetime.datetime.utcfromtimestamp(timestamp)
human = datetime.datetime.strftime(day, "%b %d %Y %H:%M:%S UTC")
return human
def process(itemid, log=True, kill=True):
if isinstance(itemid, str):
itemid = [itemid]
if isinstance(itemid, list):
if isinstance(itemid[0], str):
itemid = verify_t3(itemid)
try:
itemid = remove_existing(itemid)
temp = itemid[:]
except Exception:
return
itemid = r.get_info(thing_id=itemid)
try:
len(itemid)
except:
print(temp, "DEAD")
if kill:
logdead(temp[0])
process(temp, kill=kill)
return
for index in range(len(itemid)):
item = itemid[index]
item.idint = b36(item.id)
item.idstr = item.id
if item.distinguished is None:
item.distinguished = 0
else:
item.distinguished = DISTINGUISHMAP_R[item.distinguished]
item.url = "self" if item.is_self else item.url
item.created_utc = int(item.created_utc)
item.is_self = 1 if item.is_self else 0
item.over_18 = 1 if item.over_18 else 0
item.sub = item.subreddit.display_name
item.textlen = len(item.selftext)
try:
item.auth = item.author.name
except AttributeError:
item.auth = "[deleted]"
item = [item.idint, item.idstr, item.created_utc,
item.is_self, item.over_18, item.auth, item.title,
item.url, item.selftext, item.score, item.sub,
item.distinguished, item.textlen]
itemid[index] = item
if log:
logdb(itemid)
else:
return itemid
if len(itemid) < len(temp):
process(temp)
# 0 - idint
# 1 - idstr
# 2 - created
# 3 - self
# 4 - nsfw
# 5 - author
# 6 - title
# 7 - url
# 8 - selftext
# 9 - score
# 10 - subreddit
# 11 - distinguished
# 12 - textlen
def logdb(items):
for item in items:
cur.execute('INSERT INTO posts VALUES(?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?, ?)', item)
sql.commit()
def logdead(i):
#If an ID is dead, let's at least add it to the db.
i = i.replace('t3_', '')
data = [b36(i), i, 0, 0, 0, '?', '?', '?', '?', 0, '?', 0, 0]
logdb([data])
def verify_t3(items):
for index in range(len(items)):
i = items[index]
if 't3_' not in i:
items[index] = 't3_' + i
return items
def remove_existing(items):
done = False
items = verify_t3(items)
while not done:
done = True
for item in items:
cur.execute('SELECT * FROM posts WHERE idint=?', [b36(item[3:])])
f = cur.fetchone()
if f:
items.remove(item)
done = False
break
if len(items) == 0:
raise Exception("Nothing new")
return items
def processrange(lower, upper, kill=True):
if isinstance(lower, str):
lower = b36(lower)
if isinstance(upper, int):
upper = lower + upper
if isinstance(upper, str):
upper = b36(upper)
if upper <= lower:
print("Upper must be higher than lower")
return
ids = [b36(x) for x in range(lower, upper)]
while len(ids) > 0:
p = ids[:100]
print("%s >>> %s (%d)" % (p[0], p[-1], len(ids)))
ids = ids[100:]
process(p, kill=kill)
def lastitem():
cur.execute('SELECT * FROM posts ORDER BY idint DESC LIMIT 1')
return cur.fetchone()[1]
def show():
filea = open('show/missing.txt', 'w')
fileb = open('show/stats.txt', 'w')
cur.execute('SELECT Count(*) FROM posts')
count = cur.fetchone()
count = count[0]
counts = '{0:,}'.format(count)
mainstats = '%s posts collected; ' % counts
print('Current total:', counts)
print('Counting dead posts')
cur.execute('SELECT * FROM posts WHERE created=0')
dead = cur.fetchall()
dead = [x[1] for x in dead]
deadcount = len(dead)
deadcount = '{0:,}'.format(deadcount)
mainstats += '%s dead.\n' % deadcount
for deaditem in dead:
print(deaditem, file=filea)
filea.close()
print('Counting selfposts')
cur.execute('SELECT * FROM posts WHERE self=1')
self = cur.fetchall()
self = len(self)
link = count-self
selfs = '{0:,}'.format(self)
links = '{0:,}'.format(link)
selfstats = '%s linkposts; %s selfposts\n' % (links, selfs)
readmefile = open('README.md', 'r')
readmelines = readmefile.readlines()
readmefile.close()
readmelines[3] = mainstats
readmelines[4] = selfstats
readmefile = open('README.md', 'w')
readmefile.write(''.join(readmelines))
readmefile.close()
#STATS TIME
print('Writing subreddit stats')
cur.execute('SELECT * FROM posts')
subredditcounts = {}
while True:
fetch = cur.fetchone()
if fetch:
fetch = Post(fetch)
try:
subredditcounts[fetch.subreddit] += 1
except KeyError:
subredditcounts[fetch.subreddit] = 1
else:
break
subkeys = list(subredditcounts.keys())
subkeys.sort(key=subredditcounts.get, reverse=True)
for key in subkeys:
out = key
out += '.'*(25-len(key))
num = '{0:,}'.format(subredditcounts[key])
out += '.'*(14-len(num))
out += num
print(out, file=fileb)
fileb.close() | [
"edalool@yahoo.com"
] | edalool@yahoo.com |
3dbdb608cd2de3f1278d8f0339287fd5ce40c676 | fcc88521f63a3c22c81a9242ae3b203f2ea888fd | /Python3/0844-Backspace-String-Compare/soln-1.py | cf9b6afca02d5b7deeaed1a8aa8d927a70cbd4e0 | [
"MIT"
] | permissive | wyaadarsh/LeetCode-Solutions | b5963e3427aa547d485d3a2cb24e6cedc72804fd | 3719f5cb059eefd66b83eb8ae990652f4b7fd124 | refs/heads/master | 2022-12-06T15:50:37.930987 | 2020-08-30T15:49:27 | 2020-08-30T15:49:27 | 291,811,790 | 0 | 1 | MIT | 2020-08-31T19:57:35 | 2020-08-31T19:57:34 | null | UTF-8 | Python | false | false | 600 | py | class Solution:
def backspaceCompare(self, S, T):
"""
:type S: str
:type T: str
:rtype: bool
"""
i, j = len(S) - 1, len(T) - 1
bs, bt = 0, 0
while True:
while i >= 0 and (bs or S[i] == '#'):
bs = bs + 1 if S[i] == '#' else bs - 1
i -= 1
while j >= 0 and (bt or T[j] == '#'):
bt = bt + 1 if T[j] == '#' else bt - 1
j -= 1
if not(i >= 0 and j >= 0 and S[i] == T[j]):
return i == j == -1
i, j = i - 1, j - 1 | [
"zhang623@wisc.edu"
] | zhang623@wisc.edu |
c2527e5e639b13cb1d28b6c112b7544a225ffc45 | d538b6bae84256e241ebdc976bbe1a0d0b4ec7ee | /productDetails/forms.py | 23c107d05e7a99e435f01291b0f3565e22609a3c | [] | no_license | Sangeethsajan/WarehouseManagement | 871b3661c74922f7b524232867e5dc5d7b3dee59 | 30731a7538fa6278e9744c3060aa93f7e37fed79 | refs/heads/master | 2023-02-10T03:59:19.142042 | 2021-01-08T15:23:03 | 2021-01-08T15:23:03 | 327,938,681 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 576 | py | from django import forms
from .models import productDetails, productPath
class productForm(forms.ModelForm):
class Meta:
model = productDetails
fields ="__all__"
labels ={
'Product Name':'Name of Product',
}
def __init__(self, *args, **kwargs):
super(productForm,self).__init__(*args, **kwargs)
self.fields['category_Id'].empty_label ="Select"
self.fields['manufacturer_Id'].empty_label ="Select"
class pathForm(forms.ModelForm):
class Meta:
model = productPath
fields="__all__" | [
"31395590+Sangeethsajan@users.noreply.github.com"
] | 31395590+Sangeethsajan@users.noreply.github.com |
0f48da39bcd7577cbe908aaee672175a9921e808 | fc5e2a7873d39f93300f030eeff7678b7d09918c | /auth_middleware.py | 69158f54dc426f440f2aac2d3e5ce55a33f614c0 | [] | no_license | kubamvictor/mlflow-tracking-template | 357d01d82dd789bfab6fbda5e9ef70bdf58858c2 | 1990459979be28fb0d6495cc22ddc364f19dd2e8 | refs/heads/master | 2022-11-12T00:52:46.453922 | 2020-07-01T18:34:08 | 2020-07-01T18:34:08 | 276,577,558 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,335 | py | class AuthMiddleware():
"""
Middleware for authenticating requests sent to a flask app
"""
def __init__(self, app):
self.app = app
self.token = 'YHON36336hhngEYEY'
def __call__(self, environ, start_response):
# Validate token
if self._authenticated(environ.get('HTTP_AUTHORIZATION')):
return self.app(environ, start_response)
return self._login(environ, start_response)
def _authenticated(self, header):
"""
Function to extract and validate user token from request header
argument:
header: request header
output:
boolean: if user is validated this is set to true, otherwise false
"""
try:
if not header:
return False
_, token = header.split(None, 1)
print("User has been validated? {}".format(token == self.token))
return token == self.token
except Exception:
return False
def _login(self, environ, start_response):
start_response(
'401 Authentication Required',
[('Content-Type', 'text/html'), ('WWW-Authenticate', 'Bearer realm="mlflow server"')])
return [b'Authentication Failed! Please check your authorization token or contact system admins.']
| [
"kubamvictor@gmail.com"
] | kubamvictor@gmail.com |
41ef232e7c76144c75891a57dcc4d00868f32726 | 673f9b85708affe260b892a4eb3b1f6a0bd39d44 | /Botnets/App/App Web/PDG-env/lib/python3.6/site-packages/pandas/tests/scalar/timedelta/test_arithmetic.py | 555b47c8dc0fc9a2a6f68e9ffd73c8a00e312d4d | [
"MIT"
] | permissive | i2tResearch/Ciberseguridad_web | feee3fe299029bef96b158d173ce2d28ef1418e4 | e6cccba69335816442c515d65d9aedea9e7dc58b | refs/heads/master | 2023-07-06T00:43:51.126684 | 2023-06-26T00:53:53 | 2023-06-26T00:53:53 | 94,152,032 | 14 | 0 | MIT | 2023-09-04T02:53:29 | 2017-06-13T00:21:00 | Jupyter Notebook | UTF-8 | Python | false | false | 24,619 | py | """
Tests for scalar Timedelta arithmetic ops
"""
from datetime import datetime, timedelta
import operator
import numpy as np
import pytest
import pandas as pd
from pandas import NaT, Timedelta, Timestamp, _is_numpy_dev, offsets
import pandas._testing as tm
from pandas.core import ops
class TestTimedeltaAdditionSubtraction:
"""
Tests for Timedelta methods:
__add__, __radd__,
__sub__, __rsub__
"""
@pytest.mark.parametrize(
"ten_seconds",
[
Timedelta(10, unit="s"),
timedelta(seconds=10),
np.timedelta64(10, "s"),
np.timedelta64(10000000000, "ns"),
offsets.Second(10),
],
)
def test_td_add_sub_ten_seconds(self, ten_seconds):
# GH#6808
base = Timestamp("20130101 09:01:12.123456")
expected_add = Timestamp("20130101 09:01:22.123456")
expected_sub = Timestamp("20130101 09:01:02.123456")
result = base + ten_seconds
assert result == expected_add
result = base - ten_seconds
assert result == expected_sub
@pytest.mark.parametrize(
"one_day_ten_secs",
[
Timedelta("1 day, 00:00:10"),
Timedelta("1 days, 00:00:10"),
timedelta(days=1, seconds=10),
np.timedelta64(1, "D") + np.timedelta64(10, "s"),
offsets.Day() + offsets.Second(10),
],
)
def test_td_add_sub_one_day_ten_seconds(self, one_day_ten_secs):
# GH#6808
base = Timestamp("20130102 09:01:12.123456")
expected_add = Timestamp("20130103 09:01:22.123456")
expected_sub = Timestamp("20130101 09:01:02.123456")
result = base + one_day_ten_secs
assert result == expected_add
result = base - one_day_ten_secs
assert result == expected_sub
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_datetimelike_scalar(self, op):
# GH#19738
td = Timedelta(10, unit="d")
result = op(td, datetime(2016, 1, 1))
if op is operator.add:
# datetime + Timedelta does _not_ call Timedelta.__radd__,
# so we get a datetime back instead of a Timestamp
assert isinstance(result, Timestamp)
assert result == Timestamp(2016, 1, 11)
result = op(td, Timestamp("2018-01-12 18:09"))
assert isinstance(result, Timestamp)
assert result == Timestamp("2018-01-22 18:09")
result = op(td, np.datetime64("2018-01-12"))
assert isinstance(result, Timestamp)
assert result == Timestamp("2018-01-22")
result = op(td, NaT)
assert result is NaT
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_td(self, op):
td = Timedelta(10, unit="d")
result = op(td, Timedelta(days=10))
assert isinstance(result, Timedelta)
assert result == Timedelta(days=20)
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_pytimedelta(self, op):
td = Timedelta(10, unit="d")
result = op(td, timedelta(days=9))
assert isinstance(result, Timedelta)
assert result == Timedelta(days=19)
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_timedelta64(self, op):
td = Timedelta(10, unit="d")
result = op(td, np.timedelta64(-4, "D"))
assert isinstance(result, Timedelta)
assert result == Timedelta(days=6)
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_offset(self, op):
td = Timedelta(10, unit="d")
result = op(td, offsets.Hour(6))
assert isinstance(result, Timedelta)
assert result == Timedelta(days=10, hours=6)
def test_td_sub_td(self):
td = Timedelta(10, unit="d")
expected = Timedelta(0, unit="ns")
result = td - td
assert isinstance(result, Timedelta)
assert result == expected
def test_td_sub_pytimedelta(self):
td = Timedelta(10, unit="d")
expected = Timedelta(0, unit="ns")
result = td - td.to_pytimedelta()
assert isinstance(result, Timedelta)
assert result == expected
result = td.to_pytimedelta() - td
assert isinstance(result, Timedelta)
assert result == expected
def test_td_sub_timedelta64(self):
td = Timedelta(10, unit="d")
expected = Timedelta(0, unit="ns")
result = td - td.to_timedelta64()
assert isinstance(result, Timedelta)
assert result == expected
result = td.to_timedelta64() - td
assert isinstance(result, Timedelta)
assert result == expected
def test_td_sub_nat(self):
# In this context pd.NaT is treated as timedelta-like
td = Timedelta(10, unit="d")
result = td - NaT
assert result is NaT
def test_td_sub_td64_nat(self):
td = Timedelta(10, unit="d")
td_nat = np.timedelta64("NaT")
result = td - td_nat
assert result is NaT
result = td_nat - td
assert result is NaT
def test_td_sub_offset(self):
td = Timedelta(10, unit="d")
result = td - offsets.Hour(1)
assert isinstance(result, Timedelta)
assert result == Timedelta(239, unit="h")
def test_td_add_sub_numeric_raises(self):
td = Timedelta(10, unit="d")
for other in [2, 2.0, np.int64(2), np.float64(2)]:
with pytest.raises(TypeError):
td + other
with pytest.raises(TypeError):
other + td
with pytest.raises(TypeError):
td - other
with pytest.raises(TypeError):
other - td
def test_td_rsub_nat(self):
td = Timedelta(10, unit="d")
result = NaT - td
assert result is NaT
result = np.datetime64("NaT") - td
assert result is NaT
def test_td_rsub_offset(self):
result = offsets.Hour(1) - Timedelta(10, unit="d")
assert isinstance(result, Timedelta)
assert result == Timedelta(-239, unit="h")
def test_td_sub_timedeltalike_object_dtype_array(self):
# GH#21980
arr = np.array([Timestamp("20130101 9:01"), Timestamp("20121230 9:02")])
exp = np.array([Timestamp("20121231 9:01"), Timestamp("20121229 9:02")])
res = arr - Timedelta("1D")
tm.assert_numpy_array_equal(res, exp)
def test_td_sub_mixed_most_timedeltalike_object_dtype_array(self):
# GH#21980
now = Timestamp.now()
arr = np.array([now, Timedelta("1D"), np.timedelta64(2, "h")])
exp = np.array(
[
now - Timedelta("1D"),
Timedelta("0D"),
np.timedelta64(2, "h") - Timedelta("1D"),
]
)
res = arr - Timedelta("1D")
tm.assert_numpy_array_equal(res, exp)
def test_td_rsub_mixed_most_timedeltalike_object_dtype_array(self):
# GH#21980
now = Timestamp.now()
arr = np.array([now, Timedelta("1D"), np.timedelta64(2, "h")])
with pytest.raises(TypeError):
Timedelta("1D") - arr
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_timedeltalike_object_dtype_array(self, op):
# GH#21980
arr = np.array([Timestamp("20130101 9:01"), Timestamp("20121230 9:02")])
exp = np.array([Timestamp("20130102 9:01"), Timestamp("20121231 9:02")])
res = op(arr, Timedelta("1D"))
tm.assert_numpy_array_equal(res, exp)
@pytest.mark.parametrize("op", [operator.add, ops.radd])
def test_td_add_mixed_timedeltalike_object_dtype_array(self, op):
# GH#21980
now = Timestamp.now()
arr = np.array([now, Timedelta("1D")])
exp = np.array([now + Timedelta("1D"), Timedelta("2D")])
res = op(arr, Timedelta("1D"))
tm.assert_numpy_array_equal(res, exp)
# TODO: moved from index tests following #24365, may need de-duplication
def test_ops_ndarray(self):
td = Timedelta("1 day")
# timedelta, timedelta
other = pd.to_timedelta(["1 day"]).values
expected = pd.to_timedelta(["2 days"]).values
tm.assert_numpy_array_equal(td + other, expected)
tm.assert_numpy_array_equal(other + td, expected)
msg = r"unsupported operand type\(s\) for \+: 'Timedelta' and 'int'"
with pytest.raises(TypeError, match=msg):
td + np.array([1])
msg = r"unsupported operand type\(s\) for \+: 'numpy.ndarray' and 'Timedelta'"
with pytest.raises(TypeError, match=msg):
np.array([1]) + td
expected = pd.to_timedelta(["0 days"]).values
tm.assert_numpy_array_equal(td - other, expected)
tm.assert_numpy_array_equal(-other + td, expected)
msg = r"unsupported operand type\(s\) for -: 'Timedelta' and 'int'"
with pytest.raises(TypeError, match=msg):
td - np.array([1])
msg = r"unsupported operand type\(s\) for -: 'numpy.ndarray' and 'Timedelta'"
with pytest.raises(TypeError, match=msg):
np.array([1]) - td
expected = pd.to_timedelta(["2 days"]).values
tm.assert_numpy_array_equal(td * np.array([2]), expected)
tm.assert_numpy_array_equal(np.array([2]) * td, expected)
msg = (
"ufunc '?multiply'? cannot use operands with types"
r" dtype\('<m8\[ns\]'\) and dtype\('<m8\[ns\]'\)"
)
with pytest.raises(TypeError, match=msg):
td * other
with pytest.raises(TypeError, match=msg):
other * td
tm.assert_numpy_array_equal(td / other, np.array([1], dtype=np.float64))
tm.assert_numpy_array_equal(other / td, np.array([1], dtype=np.float64))
# timedelta, datetime
other = pd.to_datetime(["2000-01-01"]).values
expected = pd.to_datetime(["2000-01-02"]).values
tm.assert_numpy_array_equal(td + other, expected)
tm.assert_numpy_array_equal(other + td, expected)
expected = pd.to_datetime(["1999-12-31"]).values
tm.assert_numpy_array_equal(-td + other, expected)
tm.assert_numpy_array_equal(other - td, expected)
class TestTimedeltaMultiplicationDivision:
"""
Tests for Timedelta methods:
__mul__, __rmul__,
__div__, __rdiv__,
__truediv__, __rtruediv__,
__floordiv__, __rfloordiv__,
__mod__, __rmod__,
__divmod__, __rdivmod__
"""
# ---------------------------------------------------------------
# Timedelta.__mul__, __rmul__
@pytest.mark.parametrize(
"td_nat", [NaT, np.timedelta64("NaT", "ns"), np.timedelta64("NaT")]
)
@pytest.mark.parametrize("op", [operator.mul, ops.rmul])
def test_td_mul_nat(self, op, td_nat):
# GH#19819
td = Timedelta(10, unit="d")
with pytest.raises(TypeError):
op(td, td_nat)
@pytest.mark.parametrize("nan", [np.nan, np.float64("NaN"), float("nan")])
@pytest.mark.parametrize("op", [operator.mul, ops.rmul])
def test_td_mul_nan(self, op, nan):
# np.float64('NaN') has a 'dtype' attr, avoid treating as array
td = Timedelta(10, unit="d")
result = op(td, nan)
assert result is NaT
@pytest.mark.parametrize("op", [operator.mul, ops.rmul])
def test_td_mul_scalar(self, op):
# GH#19738
td = Timedelta(minutes=3)
result = op(td, 2)
assert result == Timedelta(minutes=6)
result = op(td, 1.5)
assert result == Timedelta(minutes=4, seconds=30)
assert op(td, np.nan) is NaT
assert op(-1, td).value == -1 * td.value
assert op(-1.0, td).value == -1.0 * td.value
with pytest.raises(TypeError):
# timedelta * datetime is gibberish
op(td, Timestamp(2016, 1, 2))
with pytest.raises(TypeError):
# invalid multiply with another timedelta
op(td, td)
# ---------------------------------------------------------------
# Timedelta.__div__, __truediv__
def test_td_div_timedeltalike_scalar(self):
# GH#19738
td = Timedelta(10, unit="d")
result = td / offsets.Hour(1)
assert result == 240
assert td / td == 1
assert td / np.timedelta64(60, "h") == 4
assert np.isnan(td / NaT)
def test_td_div_numeric_scalar(self):
# GH#19738
td = Timedelta(10, unit="d")
result = td / 2
assert isinstance(result, Timedelta)
assert result == Timedelta(days=5)
result = td / 5.0
assert isinstance(result, Timedelta)
assert result == Timedelta(days=2)
@pytest.mark.parametrize(
"nan",
[
np.nan,
pytest.param(
np.float64("NaN"),
marks=pytest.mark.xfail(
_is_numpy_dev,
reason="https://github.com/pandas-dev/pandas/issues/31992",
strict=False,
),
),
float("nan"),
],
)
def test_td_div_nan(self, nan):
# np.float64('NaN') has a 'dtype' attr, avoid treating as array
td = Timedelta(10, unit="d")
result = td / nan
assert result is NaT
result = td // nan
assert result is NaT
# ---------------------------------------------------------------
# Timedelta.__rdiv__
def test_td_rdiv_timedeltalike_scalar(self):
# GH#19738
td = Timedelta(10, unit="d")
result = offsets.Hour(1) / td
assert result == 1 / 240.0
assert np.timedelta64(60, "h") / td == 0.25
# ---------------------------------------------------------------
# Timedelta.__floordiv__
def test_td_floordiv_timedeltalike_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
scalar = Timedelta(hours=3, minutes=3)
assert td // scalar == 1
assert -td // scalar.to_pytimedelta() == -2
assert (2 * td) // scalar.to_timedelta64() == 2
def test_td_floordiv_null_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
assert td // np.nan is NaT
assert np.isnan(td // NaT)
assert np.isnan(td // np.timedelta64("NaT"))
def test_td_floordiv_offsets(self):
# GH#19738
td = Timedelta(hours=3, minutes=4)
assert td // offsets.Hour(1) == 3
assert td // offsets.Minute(2) == 92
def test_td_floordiv_invalid_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
with pytest.raises(TypeError):
td // np.datetime64("2016-01-01", dtype="datetime64[us]")
def test_td_floordiv_numeric_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
expected = Timedelta(hours=1, minutes=32)
assert td // 2 == expected
assert td // 2.0 == expected
assert td // np.float64(2.0) == expected
assert td // np.int32(2.0) == expected
assert td // np.uint8(2.0) == expected
def test_td_floordiv_timedeltalike_array(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
scalar = Timedelta(hours=3, minutes=3)
# Array-like others
assert td // np.array(scalar.to_timedelta64()) == 1
res = (3 * td) // np.array([scalar.to_timedelta64()])
expected = np.array([3], dtype=np.int64)
tm.assert_numpy_array_equal(res, expected)
res = (10 * td) // np.array([scalar.to_timedelta64(), np.timedelta64("NaT")])
expected = np.array([10, np.nan])
tm.assert_numpy_array_equal(res, expected)
def test_td_floordiv_numeric_series(self):
# GH#18846
td = Timedelta(hours=3, minutes=4)
ser = pd.Series([1], dtype=np.int64)
res = td // ser
assert res.dtype.kind == "m"
# ---------------------------------------------------------------
# Timedelta.__rfloordiv__
def test_td_rfloordiv_timedeltalike_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
scalar = Timedelta(hours=3, minutes=4)
# scalar others
# x // Timedelta is defined only for timedelta-like x. int-like,
# float-like, and date-like, in particular, should all either
# a) raise TypeError directly or
# b) return NotImplemented, following which the reversed
# operation will raise TypeError.
assert td.__rfloordiv__(scalar) == 1
assert (-td).__rfloordiv__(scalar.to_pytimedelta()) == -2
assert (2 * td).__rfloordiv__(scalar.to_timedelta64()) == 0
def test_td_rfloordiv_null_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
assert np.isnan(td.__rfloordiv__(NaT))
assert np.isnan(td.__rfloordiv__(np.timedelta64("NaT")))
def test_td_rfloordiv_offsets(self):
# GH#19738
assert offsets.Hour(1) // Timedelta(minutes=25) == 2
def test_td_rfloordiv_invalid_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
dt64 = np.datetime64("2016-01-01", "us")
with pytest.raises(TypeError):
td.__rfloordiv__(dt64)
def test_td_rfloordiv_numeric_scalar(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
assert td.__rfloordiv__(np.nan) is NotImplemented
assert td.__rfloordiv__(3.5) is NotImplemented
assert td.__rfloordiv__(2) is NotImplemented
with pytest.raises(TypeError):
td.__rfloordiv__(np.float64(2.0))
with pytest.raises(TypeError):
td.__rfloordiv__(np.uint8(9))
with pytest.raises(TypeError, match="Invalid dtype"):
# deprecated GH#19761, enforced GH#29797
td.__rfloordiv__(np.int32(2.0))
def test_td_rfloordiv_timedeltalike_array(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
scalar = Timedelta(hours=3, minutes=4)
# Array-like others
assert td.__rfloordiv__(np.array(scalar.to_timedelta64())) == 1
res = td.__rfloordiv__(np.array([(3 * scalar).to_timedelta64()]))
expected = np.array([3], dtype=np.int64)
tm.assert_numpy_array_equal(res, expected)
arr = np.array([(10 * scalar).to_timedelta64(), np.timedelta64("NaT")])
res = td.__rfloordiv__(arr)
expected = np.array([10, np.nan])
tm.assert_numpy_array_equal(res, expected)
def test_td_rfloordiv_numeric_series(self):
# GH#18846
td = Timedelta(hours=3, minutes=3)
ser = pd.Series([1], dtype=np.int64)
res = td.__rfloordiv__(ser)
assert res is NotImplemented
with pytest.raises(TypeError, match="Invalid dtype"):
# Deprecated GH#19761, enforced GH#29797
# TODO: GH-19761. Change to TypeError.
ser // td
# ----------------------------------------------------------------
# Timedelta.__mod__, __rmod__
def test_mod_timedeltalike(self):
# GH#19365
td = Timedelta(hours=37)
# Timedelta-like others
result = td % Timedelta(hours=6)
assert isinstance(result, Timedelta)
assert result == Timedelta(hours=1)
result = td % timedelta(minutes=60)
assert isinstance(result, Timedelta)
assert result == Timedelta(0)
result = td % NaT
assert result is NaT
def test_mod_timedelta64_nat(self):
# GH#19365
td = Timedelta(hours=37)
result = td % np.timedelta64("NaT", "ns")
assert result is NaT
def test_mod_timedelta64(self):
# GH#19365
td = Timedelta(hours=37)
result = td % np.timedelta64(2, "h")
assert isinstance(result, Timedelta)
assert result == Timedelta(hours=1)
def test_mod_offset(self):
# GH#19365
td = Timedelta(hours=37)
result = td % offsets.Hour(5)
assert isinstance(result, Timedelta)
assert result == Timedelta(hours=2)
def test_mod_numeric(self):
# GH#19365
td = Timedelta(hours=37)
# Numeric Others
result = td % 2
assert isinstance(result, Timedelta)
assert result == Timedelta(0)
result = td % 1e12
assert isinstance(result, Timedelta)
assert result == Timedelta(minutes=3, seconds=20)
result = td % int(1e12)
assert isinstance(result, Timedelta)
assert result == Timedelta(minutes=3, seconds=20)
def test_mod_invalid(self):
# GH#19365
td = Timedelta(hours=37)
with pytest.raises(TypeError):
td % Timestamp("2018-01-22")
with pytest.raises(TypeError):
td % []
def test_rmod_pytimedelta(self):
# GH#19365
td = Timedelta(minutes=3)
result = timedelta(minutes=4) % td
assert isinstance(result, Timedelta)
assert result == Timedelta(minutes=1)
def test_rmod_timedelta64(self):
# GH#19365
td = Timedelta(minutes=3)
result = np.timedelta64(5, "m") % td
assert isinstance(result, Timedelta)
assert result == Timedelta(minutes=2)
def test_rmod_invalid(self):
# GH#19365
td = Timedelta(minutes=3)
with pytest.raises(TypeError):
Timestamp("2018-01-22") % td
with pytest.raises(TypeError):
15 % td
with pytest.raises(TypeError):
16.0 % td
with pytest.raises(TypeError):
np.array([22, 24]) % td
# ----------------------------------------------------------------
# Timedelta.__divmod__, __rdivmod__
def test_divmod_numeric(self):
# GH#19365
td = Timedelta(days=2, hours=6)
result = divmod(td, 53 * 3600 * 1e9)
assert result[0] == Timedelta(1, unit="ns")
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(hours=1)
assert result
result = divmod(td, np.nan)
assert result[0] is NaT
assert result[1] is NaT
def test_divmod(self):
# GH#19365
td = Timedelta(days=2, hours=6)
result = divmod(td, timedelta(days=1))
assert result[0] == 2
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(hours=6)
result = divmod(td, 54)
assert result[0] == Timedelta(hours=1)
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(0)
result = divmod(td, NaT)
assert np.isnan(result[0])
assert result[1] is NaT
def test_divmod_offset(self):
# GH#19365
td = Timedelta(days=2, hours=6)
result = divmod(td, offsets.Hour(-4))
assert result[0] == -14
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(hours=-2)
def test_divmod_invalid(self):
# GH#19365
td = Timedelta(days=2, hours=6)
with pytest.raises(TypeError):
divmod(td, Timestamp("2018-01-22"))
def test_rdivmod_pytimedelta(self):
# GH#19365
result = divmod(timedelta(days=2, hours=6), Timedelta(days=1))
assert result[0] == 2
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(hours=6)
def test_rdivmod_offset(self):
result = divmod(offsets.Hour(54), Timedelta(hours=-4))
assert result[0] == -14
assert isinstance(result[1], Timedelta)
assert result[1] == Timedelta(hours=-2)
def test_rdivmod_invalid(self):
# GH#19365
td = Timedelta(minutes=3)
with pytest.raises(TypeError):
divmod(Timestamp("2018-01-22"), td)
with pytest.raises(TypeError):
divmod(15, td)
with pytest.raises(TypeError):
divmod(16.0, td)
with pytest.raises(TypeError):
divmod(np.array([22, 24]), td)
# ----------------------------------------------------------------
@pytest.mark.parametrize(
"op", [operator.mul, ops.rmul, operator.truediv, ops.rdiv, ops.rsub]
)
@pytest.mark.parametrize(
"arr",
[
np.array([Timestamp("20130101 9:01"), Timestamp("20121230 9:02")]),
np.array([Timestamp.now(), Timedelta("1D")]),
],
)
def test_td_op_timedelta_timedeltalike_array(self, op, arr):
with pytest.raises(TypeError):
op(arr, Timedelta("1D"))
| [
"ulcamilo@gmail.com"
] | ulcamilo@gmail.com |
13c52a06183456edede58129114517b459640d65 | 879e9f1d5ef7a48a33dc7851c7ebf8bddf8cfc2b | /apistar/backends/django_orm.py | f2ddd2f99439b61c8b22148bbced29eaa47f06a9 | [
"BSD-3-Clause"
] | permissive | mehrdadmhd/apistar | ed3dcd1f1a86ddfc602421f5f720009a535cd841 | 5cd30d380457b870961c32fad394745778d08a1c | refs/heads/master | 2021-04-09T16:04:22.655897 | 2018-03-18T11:09:40 | 2018-03-18T11:09:40 | 125,716,342 | 0 | 0 | BSD-3-Clause | 2018-03-18T11:06:28 | 2018-03-18T11:06:28 | null | UTF-8 | Python | false | false | 2,275 | py | import contextlib
import sys
import typing
import django
from django.apps import apps
from django.conf import settings as django_settings
from django.core.management import call_command
from django.db import connections, transaction
from apistar import Command, Component, Settings
class DjangoORM(object):
def __init__(self, settings: Settings) -> None:
config = {
'INSTALLED_APPS': settings.get('INSTALLED_APPS', []),
'DATABASES': settings.get('DATABASES', {}),
'AUTH_USER_MODEL': settings.get('AUTH_USER_MODEL', 'auth.User')
}
django_settings.configure(**config)
django.setup()
self.models = {
model.__name__: model
for model in apps.get_models()
}
class Session(object):
def __init__(self, orm: DjangoORM) -> None:
for name, model in orm.models.items():
setattr(self, name, model)
@contextlib.contextmanager
def get_session(backend: DjangoORM) -> typing.Generator[Session, None, None]:
"""
Create a new context-managed database session, which automatically
handles atomic rollback or commit behavior.
Args:
backend: The configured database backend.
"""
for conn in connections.all():
conn.queries_log.clear()
conn.close_if_unusable_or_obsolete()
atomic = transaction.Atomic(using=None, savepoint=True)
atomic.__enter__()
try:
yield Session(backend)
except:
exc_type, exc_value, exc_traceback = sys.exc_info()
atomic.__exit__(exc_type, exc_value, exc_traceback)
raise
exc_type, exc_value, exc_traceback = (None, None, None)
atomic.__exit__(exc_type, exc_value, exc_traceback)
def flush(): # pragma: nocover
call_command('flush', '--no-input')
def makemigrations(): # pragma: nocover
call_command('makemigrations')
def migrate():
call_command('migrate')
def showmigrations(): # pragma: nocover
call_command('showmigrations')
components = [
Component(DjangoORM),
Component(Session, init=get_session, preload=False)
]
commands = [
Command('flush', flush),
Command('makemigrations', makemigrations),
Command('migrate', migrate),
Command('showmigrations', showmigrations)
]
| [
"noreply@github.com"
] | noreply@github.com |
79697ef0b2a734acad1ab51ba409899d1f512b1f | ad314542195e2eb8ae2ba343a47a27a530f5969d | /utils/label_plot.py | 5202c7398090e2de686bc1961839e000b2d73a44 | [] | no_license | vivianyaya/SHL2021 | 183ad722f8aa34dab93c4fb5cb07f7a1f1473a35 | d610ba2b2134a915e617c81d5e24c31acac3b3c3 | refs/heads/main | 2023-06-04T03:10:05.535868 | 2021-06-25T03:19:10 | 2021-06-25T03:19:10 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,136 | py | import matplotlib.pyplot as plt
import matplotlib as mpl
import numpy as np
import seaborn as sns
plt.style.use("dark_background")
label_dic = {1: 'Still', 2: 'Walking', 3: 'Run', 4: 'Bike', 5: 'Car', 6: 'Bus', 7: 'Train', 8: 'Subway'}
cmap = mpl.cm.get_cmap('viridis', 8)
label_cols = cmap(np.linspace(0, 1, 8))
label_cols = [mpl.colors.to_hex(i) for i in label_cols]
def plot_label_each(df, col_name, this_label):
label_index = np.where(df.label == this_label)[0]
p = plt.scatter(df.index[label_index], df[col_name][label_index], c = label_cols[this_label - 1], label = label_dic[this_label])
return p
def plot_label(df, col_name):
plt.figure(figsize = [12, 8])
for this_label in np.unique(list(df.label)):
plot_label_each(df, col_name, this_label)
plt.legend(loc = 'best')
def box_label(df, col_name, use_log = False, violin = False):
plt.figure(figsize = [12, 8])
y = np.log(df[col_name]+1) if use_log else df[col_name]
if violin:
sns.violinplot(x = df['label'], y = y)
else:
sns.boxplot(x = df['label'], y = y)
plt.xticks(range(8), list(label_dic.values())) | [
"1609532549@qq.com"
] | 1609532549@qq.com |
74c507fea182bc755c7aeb7685aba03bf3b95d65 | 29592ead4965e1152a1558351b3c60005438c85f | /v0.97/dc/models/batches.py | 2969e64c06b0fa446afe5c7ef5bdd6d41c5d9319 | [
"MIT"
] | permissive | iaiting/Flask-and-pywebview-followup-application-gui | 5e489eda889c6d02ffb8efd40098a380dedd0558 | b665334403b4a8471b5f28054ee2dc7adda7d9fc | refs/heads/master | 2020-04-04T11:30:57.294114 | 2018-10-29T09:14:34 | 2018-10-29T09:14:34 | 155,893,938 | 0 | 1 | MIT | 2018-11-02T16:36:19 | 2018-11-02T16:36:18 | null | UTF-8 | Python | false | false | 13,049 | py | import os
import re
import pandas as pd
from lxml import etree
from dc.utils.commun import Commun
from dc.utils.dbwrap import Dbwrap
from dc.utils.folders import Folders
from dc.models.users import User
class Batches:
func = Commun()
conf = func.config_info()
db = Dbwrap(conf["path_to_database"])
folder = Folders()
user = User()
def process_files_paths(self, filesInfodict):
"""Extract files names, set and id for each file path"""
id_lentgh = int(self.conf["IDlentgh"])
new_opfiles_path = self.conf["path_to_new_opfiles"]
paths_to_files = list(filesInfodict.keys())
original_files_names = []
new_files_names = []
for file_path in paths_to_files:
fid = self.func.generate_id(id_lentgh)
file_name = self.func.get_file_name(file_path)
original_files_names.append(file_name)
new_file_name = os.path.join("{} {}".format(fid, file_name))
new_files_names.append(new_file_name)
files = ", ".join(original_files_names)
new_files = ", ".join(new_files_names)
return files, new_files
def get_paths_for_unassigned_prepared(self, bid_info):
"""Create the paths to move the folders from new to unassigned and prepared"""
org_path = bid_info["OriginalFilesPath"]
opfolder = "{} {}".format(bid_info["Operator"], bid_info["Aircraft"])
new_path = os.path.join(org_path, opfolder)
bid_opfolder = "{} _{}".format(opfolder, bid_info["BatchID"])
unassigned_path = os.path.join(self.conf["path_to_batches_unassigned"], bid_opfolder)
prepared_path = os.path.join(self.conf["path_to_batches_prepfiles"], bid_opfolder)
return new_path, unassigned_path, prepared_path, bid_opfolder
def add_id_to_prepfiles(self, bid_info, bid_opfolder):
"""Add the new files names to files in prepared/unassigned files"""
org_files = self.func.listify_string(bid_info["OriginalFilesName"])
new_files = self.func.listify_string(bid_info["FilesID"])
prep_path = self.conf["path_to_batches_prepfiles"]
unassg_path = self.conf["path_to_batches_unassigned"]
prep_bid_path = os.path.join(prep_path, bid_opfolder)
unassg_bid_path = os.path.join(unassg_path, bid_opfolder)
for org, new in zip(org_files, new_files):
#Paths for prepared files
preporg_path = os.path.join(prep_bid_path, org)
prepnew_path = os.path.join(prep_bid_path, new)
#Paths for unassigned files
unassgorg_path = os.path.join(unassg_bid_path, org)
unassgnew_path = os.path.join(unassg_bid_path, new)
self.func.move_folder(preporg_path, prepnew_path)
self.func.move_folder(unassgorg_path, unassgnew_path)
def copy_files_to_unassigned_prepared_dirs(self, bid_info):
"""Copy files from new folder to unassigned and prepared path folder"""
new_path, unassigned_path, prepared_path, bid_opfolder = self.get_paths_for_unassigned_prepared(bid_info)
#Raise error if a dir is found in new operator files
self.func.accept_only_files(new_path)
#Copy from new to unassigned
self.func.copy_dirs(new_path, unassigned_path)
#Copy from unassigned to prepared
self.func.copy_dirs(unassigned_path, prepared_path)
#Rename files from prepared folder
self.add_id_to_prepfiles(bid_info, bid_opfolder)
def check_file_history(self, file_info):
"""Check if an operator file was added before based on name, size and modification date"""
dfgen, conn = self.db.read_table("fileshistory", chunk_size=50000)
for df in dfgen:
df_name = df[df["FileName"] == file_info["FileName"]]
if df_name.shape[0] > 0:
df_size = df[df["FileSize"] == int(file_info["FileSize"])]
df_date = df[df["ModificationDate"] == file_info["ModificationDate"]]
if df_size.shape[0] > 0 and df_date.shape[0] > 0:
conn.close()
raise Exception("File '{}' was added before in batch '{}'".format(df_name["FileName"].tolist()[0], df_name["AddedInBatch"].tolist()[0]))
conn.close()
def prepare_fileshistory_info(self, bid_info, files_info):
"""Get rows to insert in fileshistory and check if was previously added"""
rows = []
for file_path, info in files_info["FilesInfo"].items():
file_name = self.func.get_file_name(file_path)
file_info = {"FileName": file_name,
"AddedInBatch": bid_info["BatchID"],
"FileSize": info["FileSize"],
"ModificationDate": info["ModificationDate"]}
self.check_file_history(file_info)
rows.append(file_info)
return rows
def add_batch(self):
"""Add batch to database """
bid = self.func.generate_id(int(self.conf["IDlentgh"]))
files_info = self.folder.new_opfiles_info()
files_names, new_files_names = self.process_files_paths(files_info["FilesInfo"])
userdata = self.user.session_info()
batch_info_followup = {"BatchID": bid,
"Aircraft": files_info["Aircraft"],
"Operator": files_info["Operator"],
"OriginalFilesName": files_names,
"OriginalFilesPath": self.conf["path_to_new_opfiles"],
"FilesID": new_files_names,
"AddedDate": self.func.current_date(),
"Responsible": "UNASSIGNED",
"Proofreader": "UNASSIGNED",
"ResponsibleStatus": "",
"ProofreaderStatus": "",
"OverallStatus": "UNASSIGNED",
"ChangesLog": "Batch added by {},".format(userdata["User"])}
batch_info_fileshistory = self.prepare_fileshistory_info(batch_info_followup, files_info)
#Data for db is prepared now copy the files.
self.copy_files_to_unassigned_prepared_dirs(batch_info_followup)
#Now the files a copied now insert data to database
#Insert the batch
self.db.create_row("followup", batch_info_followup)
#Insert the file history
for file_history in batch_info_fileshistory:
self.db.create_row("fileshistory", file_history)
return bid
def followup_reversed(self):
"""Get followup table from db reverse it and return it as dict"""
df = self.db.read_table("followup")
df_dict = df.reindex(index=df.index[::-1]).to_dict('list')
return df_dict
def followup_for_responsible(self):
"""Get followup table from db reverse it filter it for current user and return it as dict"""
userdata = self.user.session_info()
df = self.db.select_row("followup", "Responsible", userdata["User"], asdict=False)
df_dict = df.reindex(index=df.index[::-1]).to_dict('list')
return df_dict
def get_batch(self, bid=""):
"""Get batch id if specified"""
bid_data = self.db.select_row("followup", "BatchID", bid)
data = {}
for col, val in bid_data.items():
data[col] = val[0]
return data
def bid_options(self):
"""Batch status options from config.json file"""
batch_opt = {
"status_user" : self.func.listify_string(self.conf["batch_status_options_responsible"]),
"status_proofreader" : self.func.listify_string(self.conf["batch_status_options_proofreader"]),
"status_overall" : self.func.listify_string(self.conf["batch_status_options_overall"]),
"aircrafts" : self.func.listify_string(self.conf["aircrafts"]),
"split_batch_factor" : self.func.listify_string(self.conf["split_batch_factor"]),
}
bid_data = self.func.read_json("batch.json")
batch_opt.update(bid_data)
batch_opt.update(self.user.context_disable())
batch_opt.update({"users": self.user.get_users(), "proofreaders": self.user.get_proofreaders()})
return batch_opt
def set_default_proofreader(self, data):
"""Update with the default proofreader for the given responsible if needed"""
if "Responsible" in list(data.keys()):
user_data = self.db.select_row("users", "User", data["Responsible"])
data["Proofreader"] = user_data["Proofreader"][0]
data["ResponsibleStatus"] = ""
data["ProofreaderStatus"] = ""
data["OverallStatus"] = "ONGOING"
data["StartDate"] = self.func.current_date()
bid_info = self.func.read_json("batch.json")
data["ChangesLog"] = bid_info["ChangesLog"] + "\nAssigned to {} on {},".format(data["Responsible"], data["StartDate"])
return data
else:
return data
def clear_start_date(self, data):
"""If Status is STANDBY or UNRECORDABLE then clear start date"""
if "ProofreaderStatus" in list(data.keys()):
if data["ProofreaderStatus"] == "UNRECORDABLE" or data["ProofreaderStatus"] == "STANDBY":
data["StartDate"] = ""
data["ImportedDateISAIM"] = ""
data["TotalRowsNbr"] = ""
data["MPDTaskRowsNbr"] = ""
data["OperatorRowsNbr"] = ""
data["FindingsRowsNbr"] = ""
data["EstimatedTaskNbr"] = ""
data["EstimatedFdgNbr"] = ""
return data
def dcs_info(self, dcspath):
"""Get xml info from the extracted xml from FE"""
file = open(dcspath)
tree = etree.parse(file)
sumAll = tree.xpath('//sum')
totalRows = sum([int(s.text) for s in sumAll])
sumMpd = tree.xpath('//mpdTask//sum')
mpdtask = sum([int(s.text) for s in sumMpd])
sumOp = tree.xpath('//opeTask//sum')
optask = sum([int(s.text) for s in sumOp])
sumFindings = tree.xpath("//finding[@activated='true']//sum")
findings = sum([int(s.text) for s in sumFindings])
info_dcs = {"TotalRowsNbr": totalRows,
"MPDTaskRowsNbr": mpdtask,
"OperatorRowsNbr": optask,
"FindingsRowsNbr": findings
}
return info_dcs
def update_ifstatus_finished(self, data):
"""If Status is FINISHED update ImportedDateISAIM column"""
if "ProofreaderStatus" in list(data.keys()):
if data["ProofreaderStatus"] == "FINISHED":
data["ImportedDateISAIM"] = self.func.current_date()
return data
def update_ifstatus_toimport(self, data):
"""If ProofreaderStatus is To import then look for dcs info and update data"""
if "ProofreaderStatus" in list(data.keys()):
if data["ProofreaderStatus"] == "TO BE IMPORTED":
dcs_files = os.listdir(self.conf["path_to_dcs_info"])
dcs_file = [f for f in dcs_files if data["BatchID"] in f]
if len(dcs_file) == 1:
dcs_file_path = os.path.join(self.conf["path_to_dcs_info"], dcs_file[0])
dcs_result = self.dcs_info(dcs_file_path)
data.update(dcs_result)
return data
else:
raise Exception("DCS file for batch '{}' not found!".format(data["BatchID"]))
else:
return data
def process_status_batch_form(self, data):
"""Process dict from, form received from update_status page"""
data = self.func.remove_null(data)
data = self.set_default_proofreader(data)
data = self.clear_start_date(data)
data = self.update_ifstatus_toimport(data)
data = self.update_ifstatus_finished(data)
#Move if needed the folders
self.folder.move_prepfile_in_assigned(data) #assign
self.folder.move_assigned_in_tbchecked(data) #to be checked
self.folder.move_tobchecked_in_tbimported(data) #to be imported
self.folder.move_tbimported_in_finished(data) #finished
self.folder.move_tbchecked_in_assigned(data) #rework
self.folder.move_tobchecked_in_standby(data) #standby
self.folder.move_tobchecked_in_unrecordable(data) #unrecordable
return self.db.update_row("followup", data, "BatchID") | [
"noreply@github.com"
] | noreply@github.com |
1dd9830ae74fa5d06a572e1f0b7f6445fd3ae66c | bb27630e7af7f1bccbb5cfc892d0b0e6445fe874 | /05_django/01_djangoIntro/django_formTest/formTestApp/apps.py | f318035ab1e56dfebb9f0f7f047dd9a3a89e5c53 | [] | no_license | MrBreakIT/pythonStack | fe8cd9418ee1060ada8cd1c446332d81facecf4e | f615436dbff581c50ded70dec6532f6339977c1d | refs/heads/main | 2023-02-06T14:37:05.688885 | 2020-12-27T22:37:09 | 2020-12-27T22:37:09 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 97 | py | from django.apps import AppConfig
class FormtestAppConfig(AppConfig):
name = 'formTestApp'
| [
"johnpike1022@gmail.com"
] | johnpike1022@gmail.com |
829fb7ef3ac2065f715fbf5de042422456c84678 | 0fba7a95a16bd8351ccfdf806dde5f48f1e0b137 | /localsite/migrations/0031_auto__add_field_seatlocation_hallscheme.py | 2af37f723707ba51878ebeb9ab87092cba79eda7 | [] | no_license | yanchenko-igor/shop-ticket | 3e36c2c664241391fc02e90d239ff9458e601209 | 0093a3c1eb8025f480ecf67692ad1340ddf38296 | refs/heads/master | 2020-05-21T12:13:18.652894 | 2011-12-02T15:00:44 | 2011-12-02T15:00:44 | 3,631,970 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 14,138 | py | # encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Adding field 'SeatLocation.hallscheme'
db.add_column('localsite_seatlocation', 'hallscheme', self.gf('django.db.models.fields.related.ForeignKey')(default=1, related_name='seats', to=orm['localsite.HallScheme']), keep_default=False)
def backwards(self, orm):
# Deleting field 'SeatLocation.hallscheme'
db.delete_column('localsite_seatlocation', 'hallscheme_id')
models = {
'localsite.announcement': {
'Meta': {'ordering': "['ordering', 'begin']", 'object_name': 'Announcement'},
'begin': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}),
'end': ('django.db.models.fields.DateTimeField', [], {'null': 'True', 'blank': 'True'}),
'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'announcements'", 'to': "orm['localsite.Event']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'ordering': ('django.db.models.fields.IntegerField', [], {'default': '0'}),
'text': ('django.db.models.fields.TextField', [], {})
},
'localsite.city': {
'Meta': {'ordering': "['ordering', 'name']", 'object_name': 'City'},
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '25'}),
'ordering': ('django.db.models.fields.IntegerField', [], {'default': '0'}),
'slug': ('django.db.models.fields.SlugField', [], {'max_length': '25', 'db_index': 'True'})
},
'localsite.event': {
'Meta': {'ordering': "['min_date', 'max_date']", 'object_name': 'Event'},
'hallscheme': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'events'", 'to': "orm['localsite.HallScheme']"}),
'max_date': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}),
'max_price': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}),
'min_date': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}),
'min_price': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}),
'product': ('django.db.models.fields.related.OneToOneField', [], {'to': "orm['product.Product']", 'unique': 'True', 'primary_key': 'True'}),
'tags': ('tagging.fields.TagField', [], {})
},
'localsite.eventdate': {
'Meta': {'ordering': "['datetime', 'event']", 'unique_together': "(('event', 'datetime'),)", 'object_name': 'EventDate'},
'datetime': ('django.db.models.fields.DateTimeField', [], {}),
'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'dates'", 'to': "orm['localsite.Event']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'map': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'})
},
'localsite.hall': {
'Meta': {'ordering': "['city', 'name']", 'unique_together': "(('name', 'city'),)", 'object_name': 'Hall'},
'address': ('django.db.models.fields.CharField', [], {'max_length': '255', 'blank': 'True'}),
'city': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'halls'", 'to': "orm['localsite.City']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '120'})
},
'localsite.hallscheme': {
'Meta': {'ordering': "['hall', 'name']", 'unique_together': "(('name', 'hall'),)", 'object_name': 'HallScheme'},
'hall': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'schemes'", 'to': "orm['localsite.Hall']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'map': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '75'}),
'substrate': ('django.db.models.fields.files.FileField', [], {'max_length': '100'})
},
'localsite.seatgroup': {
'Meta': {'ordering': "['hallscheme', 'name']", 'unique_together': "(('hallscheme', 'name'), ('hallscheme', 'slug'))", 'object_name': 'SeatGroup'},
'hallscheme': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'groups'", 'to': "orm['localsite.HallScheme']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '25'}),
'slug': ('django.db.models.fields.SlugField', [], {'db_index': 'True', 'max_length': '50', 'blank': 'True'})
},
'localsite.seatgroupprice': {
'Meta': {'ordering': "['event', 'group']", 'unique_together': "(('group', 'event'),)", 'object_name': 'SeatGroupPrice'},
'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'prices'", 'to': "orm['localsite.Event']"}),
'group': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'prices'", 'to': "orm['localsite.SeatGroup']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'price': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'})
},
'localsite.seatlocation': {
'Meta': {'ordering': "['section', 'row', 'col']", 'unique_together': "(('section', 'row', 'col'),)", 'object_name': 'SeatLocation'},
'col': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}),
'group': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'seats'", 'to': "orm['localsite.SeatGroup']"}),
'hallscheme': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'seats'", 'to': "orm['localsite.HallScheme']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'row': ('django.db.models.fields.IntegerField', [], {'null': 'True', 'blank': 'True'}),
'section': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'seats'", 'to': "orm['localsite.SeatSection']"}),
'slug': ('django.db.models.fields.SlugField', [], {'db_index': 'True', 'max_length': '50', 'blank': 'True'})
},
'localsite.seatsection': {
'Meta': {'ordering': "['hallscheme', 'name']", 'unique_together': "(('hallscheme', 'name'),)", 'object_name': 'SeatSection'},
'hallscheme': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'sections'", 'to': "orm['localsite.HallScheme']"}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '25'}),
'slug': ('django.db.models.fields.SlugField', [], {'db_index': 'True', 'max_length': '50', 'blank': 'True'})
},
'localsite.ticket': {
'Meta': {'ordering': "['datetime', 'event']", 'unique_together': "(('product', 'datetime', 'seat'), ('event', 'datetime', 'seat'))", 'object_name': 'Ticket'},
'datetime': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'tickets'", 'to': "orm['localsite.EventDate']"}),
'event': ('django.db.models.fields.related.ForeignKey', [], {'related_name': "'tickets'", 'to': "orm['localsite.Event']"}),
'product': ('django.db.models.fields.related.OneToOneField', [], {'to': "orm['product.Product']", 'unique': 'True', 'primary_key': 'True'}),
'seat': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['localsite.SeatLocation']"}),
'status': ('django.db.models.fields.CharField', [], {'default': "'freely'", 'max_length': '8'})
},
'product.category': {
'Meta': {'ordering': "['site', 'parent__id', 'ordering', 'name']", 'unique_together': "(('site', 'slug'),)", 'object_name': 'Category'},
'description': ('django.db.models.fields.TextField', [], {'blank': 'True'}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}),
'meta': ('django.db.models.fields.TextField', [], {'null': 'True', 'blank': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '200'}),
'ordering': ('django.db.models.fields.IntegerField', [], {'default': '0'}),
'parent': ('django.db.models.fields.related.ForeignKey', [], {'blank': 'True', 'related_name': "'child'", 'null': 'True', 'to': "orm['product.Category']"}),
'related_categories': ('django.db.models.fields.related.ManyToManyField', [], {'blank': 'True', 'related_name': "'related_categories_rel_+'", 'null': 'True', 'to': "orm['product.Category']"}),
'site': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['sites.Site']"}),
'slug': ('django.db.models.fields.SlugField', [], {'db_index': 'True', 'max_length': '50', 'blank': 'True'})
},
'product.product': {
'Meta': {'ordering': "('site', 'ordering', 'name')", 'unique_together': "(('site', 'sku'), ('site', 'slug'))", 'object_name': 'Product'},
'active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}),
'also_purchased': ('django.db.models.fields.related.ManyToManyField', [], {'blank': 'True', 'related_name': "'also_purchased_rel_+'", 'null': 'True', 'to': "orm['product.Product']"}),
'category': ('django.db.models.fields.related.ManyToManyField', [], {'to': "orm['product.Category']", 'symmetrical': 'False', 'blank': 'True'}),
'date_added': ('django.db.models.fields.DateField', [], {'null': 'True', 'blank': 'True'}),
'description': ('django.db.models.fields.TextField', [], {'default': "''", 'blank': 'True'}),
'featured': ('django.db.models.fields.BooleanField', [], {'default': 'False'}),
'height': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '6', 'decimal_places': '2', 'blank': 'True'}),
'height_units': ('django.db.models.fields.CharField', [], {'max_length': '3', 'null': 'True', 'blank': 'True'}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'items_in_stock': ('django.db.models.fields.DecimalField', [], {'default': "'0'", 'max_digits': '18', 'decimal_places': '6'}),
'length': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '6', 'decimal_places': '2', 'blank': 'True'}),
'length_units': ('django.db.models.fields.CharField', [], {'max_length': '3', 'null': 'True', 'blank': 'True'}),
'meta': ('django.db.models.fields.TextField', [], {'max_length': '200', 'null': 'True', 'blank': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '255'}),
'ordering': ('django.db.models.fields.IntegerField', [], {'default': '0'}),
'related_items': ('django.db.models.fields.related.ManyToManyField', [], {'blank': 'True', 'related_name': "'related_items_rel_+'", 'null': 'True', 'to': "orm['product.Product']"}),
'shipclass': ('django.db.models.fields.CharField', [], {'default': "'DEFAULT'", 'max_length': '10'}),
'short_description': ('django.db.models.fields.TextField', [], {'default': "''", 'max_length': '200', 'blank': 'True'}),
'site': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['sites.Site']"}),
'sku': ('django.db.models.fields.CharField', [], {'max_length': '255', 'null': 'True', 'blank': 'True'}),
'slug': ('django.db.models.fields.SlugField', [], {'db_index': 'True', 'max_length': '255', 'blank': 'True'}),
'taxClass': ('django.db.models.fields.related.ForeignKey', [], {'to': "orm['product.TaxClass']", 'null': 'True', 'blank': 'True'}),
'taxable': ('django.db.models.fields.BooleanField', [], {'default': 'False'}),
'total_sold': ('django.db.models.fields.DecimalField', [], {'default': "'0'", 'max_digits': '18', 'decimal_places': '6'}),
'weight': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '8', 'decimal_places': '2', 'blank': 'True'}),
'weight_units': ('django.db.models.fields.CharField', [], {'max_length': '3', 'null': 'True', 'blank': 'True'}),
'width': ('django.db.models.fields.DecimalField', [], {'null': 'True', 'max_digits': '6', 'decimal_places': '2', 'blank': 'True'}),
'width_units': ('django.db.models.fields.CharField', [], {'max_length': '3', 'null': 'True', 'blank': 'True'})
},
'product.taxclass': {
'Meta': {'object_name': 'TaxClass'},
'description': ('django.db.models.fields.CharField', [], {'max_length': '30'}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'title': ('django.db.models.fields.CharField', [], {'max_length': '20'})
},
'sites.site': {
'Meta': {'ordering': "('domain',)", 'object_name': 'Site', 'db_table': "'django_site'"},
'domain': ('django.db.models.fields.CharField', [], {'max_length': '100'}),
'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}),
'name': ('django.db.models.fields.CharField', [], {'max_length': '50'})
}
}
complete_apps = ['localsite']
| [
"yanchenko.igor@gmail.com"
] | yanchenko.igor@gmail.com |
4b26d689090b162798e9ad7184c3ae1a58f20af4 | 67113383dcc2a9484a82cfc6de3e1546d1000a13 | /accounts/filter.py | da704192a0f60f8d439fdbe3b7eee7521b809bba | [] | no_license | Yair-BD/Customer-Managment | 4ef703f6562b88875086425d8509ba8aa363c0cd | a92fb7341d5383da2e8f15818133b955174ce05b | refs/heads/master | 2023-06-26T07:16:41.855959 | 2021-07-27T11:48:06 | 2021-07-27T11:48:06 | 382,893,936 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 578 | py | import django_filters
from .models import *
from django_filters import DateFilter, CharFilter
class OrderFilter(django_filters.FilterSet):
start_date = DateFilter(field_name='date_create', lookup_expr='gte')
end_date = DateFilter(field_name='date_create', lookup_expr='lte')
note = CharFilter(field_name='note', lookup_expr='icontains') # נון את היכולת לחפש ביותר מדוייק
class Meta:
model = Order
fields = '__all__'
exclude = ['customer', 'date_create', 'date_update' ] # ברירת משתתפים בחיפוש | [
"79565335+Yair-BD@users.noreply.github.com"
] | 79565335+Yair-BD@users.noreply.github.com |
d62cb9ed15fdf25fbcf76191f5229784b9ee13e5 | cf3ef8f3eca858bd3c64ba6159a2ba7cdb1722ad | /studygroups/views/organizer.py | 06e748fc35dd2c27b9b26843e6b0676e4ae6d0d6 | [] | no_license | alvarmaciel/learning-circles | 2ff956dcbe0b5a42f64036c33613644115063a8d | 3ac444fd6f5a81f655face733e7d41786e085cd4 | refs/heads/master | 2021-01-11T00:45:04.513019 | 2016-10-05T14:13:16 | 2016-10-05T14:13:16 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,094 | py | import datetime
from django.shortcuts import render, render_to_response, get_object_or_404
from django.template import RequestContext
from django.template.loader import render_to_string
from django.core.urlresolvers import reverse, reverse_lazy
from django.contrib.auth.decorators import login_required
from django.contrib.auth.models import User
from django.core.exceptions import PermissionDenied
from django.core.mail import EmailMultiAlternatives, send_mail
from django.contrib import messages
from django.conf import settings
from django import http
from django.views.decorators.http import require_http_methods
from django.views.decorators.csrf import csrf_exempt
from django.utils import timezone
from django.utils.translation import ugettext as _
from django.views.generic.base import View
from django.views.generic.edit import CreateView, UpdateView, DeleteView
from django.views.generic import ListView
from studygroups.models import Course
from studygroups.models import StudyGroup
from studygroups.models import TeamMembership
from studygroups.models import Facilitator
from studygroups.models import StudyGroupMeeting
from studygroups.models import report_data
from studygroups.models import generate_all_meetings
from studygroups.models import get_team_users
from studygroups.models import get_user_team
from studygroups.forms import StudyGroupForm
from studygroups.forms import FacilitatorForm
from studygroups.decorators import user_is_organizer
@user_is_organizer
def organize(request):
today = datetime.datetime.now().date()
two_weeks_ago = today - datetime.timedelta(weeks=2, days=today.weekday())
two_weeks = today - datetime.timedelta(days=today.weekday()) + datetime.timedelta(weeks=3)
study_groups = StudyGroup.objects.active()
facilitators = Facilitator.objects.all()
courses = []# TODO Remove courses until we implement course selection for teams
team = None
if not request.user.is_staff:
team = get_user_team(request.user)
team_users = get_team_users(request.user)
study_groups = study_groups.filter(facilitator__in=team_users)
facilitators = facilitators.filter(user__in=team_users)
active_study_groups = study_groups.filter(
id__in=StudyGroupMeeting.objects.active().filter(meeting_date__gte=two_weeks_ago).values('study_group')
)
meetings = StudyGroupMeeting.objects.active()\
.filter(study_group__in=study_groups, meeting_date__gte=two_weeks_ago)\
.exclude(meeting_date__gte=two_weeks)
context = {
'team': team,
'courses': courses,
'meetings': meetings,
'study_groups': study_groups,
'active_study_groups': active_study_groups,
'facilitators': facilitators,
'today': timezone.now(),
}
return render_to_response('studygroups/organize.html', context, context_instance=RequestContext(request))
class StudyGroupList(ListView):
model = StudyGroup
def get_queryset(self):
study_groups = StudyGroup.objects.active()
if not self.request.user.is_staff:
team_users = get_team_users(self.request.user)
study_groups = study_groups.filter(facilitator__in=team_users)
return study_groups
class StudyGroupMeetingList(ListView):
model = StudyGroupMeeting
def get_queryset(self):
study_groups = StudyGroup.objects.active()
if not self.request.user.is_staff:
team_users = get_team_users(self.request.user)
study_groups = study_groups.filter(facilitator__in=team_users)
meetings = StudyGroupMeeting.objects.active().filter(study_group__in=study_groups)
return meetings
class TeamMembershipDelete(DeleteView):
model = TeamMembership
success_url = reverse_lazy('studygroups_organize')
template_name = 'studygroups/confirm_delete_membership.html'
def get_object(self, queryset=None):
if queryset == None:
queryset = TeamMembership.objects
return queryset.get(user_id=self.kwargs.get('user_id'), team_id=self.kwargs.get('team_id'))
class CourseUpdate(UpdateView):
model = Course
fields = [
'title',
'provider',
'link',
'start_date',
'duration',
'prerequisite',
'time_required',
'caption',
]
success_url = reverse_lazy('studygroups_organize')
class CourseDelete(DeleteView):
model = Course
success_url = reverse_lazy('studygroups_organize')
template_name = 'studygroups/confirm_delete.html'
class StudyGroupCreate(CreateView):
model = StudyGroup
form_class = StudyGroupForm
success_url = reverse_lazy('studygroups_organize')
def form_valid(self, form):
self.object = form.save()
generate_all_meetings(self.object)
return http.HttpResponseRedirect(self.get_success_url())
@user_is_organizer
def report(request):
# TODO - remove this view
study_groups = StudyGroup.objects.active()
for study_group in study_groups:
study_group.laptop_stats = {}
context = {
'study_groups': study_groups,
}
return render_to_response('studygroups/report.html', context, context_instance=RequestContext(request))
@user_is_organizer
def weekly_report(request, year=None, month=None, day=None ):
today = timezone.now().replace(hour=0, minute=0, second=0, microsecond=0)
if month and day and year:
today = today.replace(year=int(year), month=int(month), day=int(day))
start_time = today - datetime.timedelta(days=today.weekday())
end_time = start_time + datetime.timedelta(days=7)
context = {
'start_time': start_time,
'end_time': end_time,
}
# get team for current user
team = None
membership = TeamMembership.objects.filter(user=request.user, role=TeamMembership.ORGANIZER).first()
if membership:
team = membership.team
context.update(report_data(start_time, end_time, team))
return render_to_response('studygroups/weekly-update.html', context, context_instance=RequestContext(request))
| [
"dirkcuys@gmail.com"
] | dirkcuys@gmail.com |
148610beec78d8585a9828b6d141e4d2459acfd4 | 77f47db313b4f018d64f0f5095ac10ae12655507 | /campaign/urls.py | 48cbb5118bac3376b0b1526b54854552694d9177 | [] | no_license | idlivada/campaign | e03ef58cb9fac05513734625ce2b94d4a430da4f | d85ad505c420b71ca4bf2de754b53e8ec0d7629e | refs/heads/master | 2020-04-16T16:40:49.666387 | 2017-06-27T12:49:58 | 2017-06-27T12:49:58 | 19,208,632 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 663 | py | from django.conf.urls import patterns, include, url
from django.conf.urls.static import static
from django.conf import settings
from django.contrib import admin
admin.autodiscover()
urlpatterns = patterns('',
url(r'^admin/', include(admin.site.urls)),
url(r'^^$', 'core.views.home', name='home'),
url(r'^locator/', 'core.views.locator', name='locator'),
url(r'^call/', 'core.views.call', name='call'),
url(r'^dial-callback/', 'core.views.dial_callback', name='dial-callback'),
url(r'^s/(?P<slug>[a-zA-Z0-9_.-]+)/', 'core.views.single_campaign', name='single-campaign'),
) + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
| [
"idlivada@gmail.com"
] | idlivada@gmail.com |
bb146c14c39b06484ad348ceeed461644b34d13c | df3853b41ed05d86f5bcd992fcc265f637c67784 | /graph/blini1.py | a90284faa70677f0bfa9cbcff1a51bda97d6ec7c | [] | no_license | KseniaMIPT/Adamasta | 6ab0121519581dbbbf6ae788d1da85f545f718d1 | e91c34c80834c3f4bf176bc4bf6bf790f9f72ca3 | refs/heads/master | 2021-01-10T16:48:31.141709 | 2016-11-23T21:02:25 | 2016-11-23T21:02:25 | 43,350,507 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 100 | py | def check():
for i in range(5):
yield i+1
print(check())
for i in check():
print(i)
| [
"ksenia22.11@yandex.ru"
] | ksenia22.11@yandex.ru |
219af0b72ccdda85c00c48773350f2e417059536 | a9d079617071855dfbd183cac5a9588ab6e30d84 | /models/test_model.py | 791316b6416c23091bf5fa7b3878ea7ab5ee4ec2 | [
"BSD-3-Clause",
"BSD-2-Clause"
] | permissive | yunyanxing/pairwise_xray_augmentation | 1ecb1a1114688b307673e5346a27f6300dd2874c | 057927cf247b3854e311fbdf52b74001175915d8 | refs/heads/master | 2020-08-09T13:32:21.899166 | 2020-01-22T00:58:49 | 2020-01-22T00:58:49 | 214,098,075 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,726 | py | from .base_model import BaseModel
from . import networks
from .cycle_gan_model import CycleGANModel
class TestModel(BaseModel):
def name(self):
return 'TestModel'
@staticmethod
def modify_commandline_options(parser, is_train=True):
assert not is_train, 'TestModel cannot be used in train mode'
parser = CycleGANModel.modify_commandline_options(parser, is_train=False)
parser.set_defaults(dataset_mode='single')
parser.add_argument('--model_suffix', type=str, default='',
help='In checkpoints_dir, [epoch]_net_G[model_suffix].pth will'
' be loaded as the generator of TestModel')
return parser
def initialize(self, opt):
assert (not opt.isTrain)
BaseModel.initialize(self, opt)
# specify the training losses you want to print out. The program will call base_model.get_current_losses
self.loss_names = []
# specify the images you want to save/display. The program will call base_model.get_current_visuals
self.visual_names = ['real_A', 'fake_B']
# specify the models you want to save to the disk. The program will call base_model.save_networks and base_model.load_networks
self.model_names = ['G' + opt.model_suffix]
self.netG = networks.define_G(opt.input_nc, opt.output_nc, opt.ngf, opt.netG,
opt.norm, not opt.no_dropout, opt.init_type, opt.init_gain, self.gpu_ids)
# assigns the model to self.netG_[suffix] so that it can be loaded
# please see BaseModel.load_networks
setattr(self, 'netG' + opt.model_suffix, self.netG)
def set_input(self, input):
# we need to use single_dataset mode
self.real_A = input['A'].to(self.device)
self.image_paths = input['A_paths']
def forward(self):
self.fake_B = self.netG(self.real_A)
| [
"yunyan.xing@monash.edu"
] | yunyan.xing@monash.edu |
ebc550669115d5534bc61b8f5cd01e7b0b38ed40 | c32d6a87858139a66dde152097e333eee92a08de | /csrv/model/actions/play_event_action.py | a57f67402f82243d5b293609aa208f7bb54e38fa | [
"Apache-2.0"
] | permissive | cpjolicoeur/CentralServer | 79b8b5a3a3ed4aa29167ecdcb4e40983a1d5e598 | e377c65d8f3adf5a2d3273acd4f459be697aea56 | refs/heads/master | 2020-12-29T00:12:31.245099 | 2014-03-08T07:33:38 | 2014-03-08T07:33:38 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 823 | py | """Base actions for the players to take."""
from csrv.model.actions import action
from csrv.model import cost
from csrv.model import errors
from csrv.model import events
from csrv.model import game_object
from csrv.model import parameters
class PlayEventAction(action.Action):
DESCRIPTION = '[click]: Play an event'
COST_CLASS = cost.EventCost
def resolve(self, response=None, ignore_clicks=False, ignore_all_costs=False):
action.Action.resolve(
self,
ignore_clicks=ignore_clicks,
ignore_all_costs=ignore_all_costs)
self.player.grip.remove(self.card)
self.card.is_faceup = True
self.card.play()
self.card.trash()
self.game.log('The runner plays %s' % self.card.NAME, self.card.game_id)
@property
def description(self):
return 'Play %s' % self.card.NAME
| [
"mrroach@okmaybe.com"
] | mrroach@okmaybe.com |
a4016f8a6330e19ed8a96f4454f25d67f80f281e | b8ec3e360f091240304afcbebb0860730ac3d835 | /mpu6050.py | c18ac2caf048a629782f454ffde99e9326f409a3 | [] | no_license | brbza/motion | 89a8a9a2ede6906ecb9bf25edd9b90d2c4cd94cf | d903e346e941e5d3e96363a97eaf27fd0ba04019 | refs/heads/master | 2023-05-11T15:03:43.891032 | 2019-06-24T08:05:20 | 2019-06-24T08:05:20 | 193,440,400 | 0 | 0 | null | 2023-05-01T20:35:16 | 2019-06-24T05:35:16 | Python | UTF-8 | Python | false | false | 3,946 | py | import machine
class MPU6050(object):
SELF_TEST_X = 0x0D
SELF_TEST_Y = 0x0E
SELF_TEST_Z = 0x0F
SELF_TEST_A = 0x10
SMPLRT_DIV = 0x19
CONFIG = 0x1A
GYRO_CONFIG = 0x1B
ACCEL_CONFIG = 0x1C
FIFO_EN = 0x23
I2C_MST_CTRL = 0x24
I2C_SLV0_ADDR = 0x25
I2C_SLV0_REG = 0x26
I2C_SLV0_CTRL = 0x27
I2C_SLV1_ADDR = 0x28
I2C_SLV1_REG = 0x29
I2C_SLV1_CTRL = 0x2A
I2C_SLV2_ADDR = 0x2B
I2C_SLV2_REG = 0x2C
I2C_SLV2_CTRL = 0x2D
I2C_SLV3_ADDR = 0x2E
I2C_SLV3_REG = 0x2F
I2C_SLV3_CTRL = 0x30
I2C_SLV4_ADDR = 0x31
I2C_SLV4_REG = 0x32
I2C_SLV4_DO = 0x33
I2C_SLV4_CTRL = 0x34
I2C_SLV4_DI = 0x35
I2C_MST_STATUS = 0x36
INT_PIN_CFG = 0x37
INT_ENABLE = 0x38
INT_STATUS = 0x3A
ACCEL_XOUT_H = 0x3B
ACCEL_XOUT_L = 0x3C
ACCEL_YOUT_H = 0x3D
ACCEL_YOUT_L = 0x3E
ACCEL_ZOUT_H = 0x3F
ACCEL_ZOUT_L = 0x40
TEMP_OUT_H = 0x41
TEMP_OUT_L = 0x42
GYRO_XOUT_H = 0x43
GYRO_XOUT_L = 0x44
GYRO_YOUT_H = 0x45
GYRO_YOUT_L = 0x46
GYRO_ZOUT_H = 0x47
GYRO_ZOUT_L = 0x48
EXT_SENS_DATA_00 = 0x49
EXT_SENS_DATA_01 = 0x4A
EXT_SENS_DATA_02 = 0x4B
EXT_SENS_DATA_03 = 0x4C
EXT_SENS_DATA_04 = 0x4D
EXT_SENS_DATA_05 = 0x4E
EXT_SENS_DATA_06 = 0x4F
EXT_SENS_DATA_07 = 0x50
EXT_SENS_DATA_08 = 0x51
EXT_SENS_DATA_09 = 0x52
EXT_SENS_DATA_10 = 0x53
EXT_SENS_DATA_11 = 0x54
EXT_SENS_DATA_12 = 0x55
EXT_SENS_DATA_13 = 0x56
EXT_SENS_DATA_14 = 0x57
EXT_SENS_DATA_15 = 0x58
EXT_SENS_DATA_16 = 0x59
EXT_SENS_DATA_17 = 0x5A
EXT_SENS_DATA_18 = 0x5B
EXT_SENS_DATA_19 = 0x5C
EXT_SENS_DATA_20 = 0x5D
EXT_SENS_DATA_21 = 0x5E
EXT_SENS_DATA_22 = 0x5F
EXT_SENS_DATA_23 = 0x60
I2C_SLV0_DO = 0x63
I2C_SLV1_DO = 0x64
I2C_SLV2_DO = 0x65
I2C_SLV3_DO = 0x66
I2C_MST_DELAY_CTRL = 0x67
SIGNAL_PATH_RESET = 0x68
USER_CTRL = 0x6A
PWR_MGMT_1 = 0x6B
PWR_MGMT_2 = 0x6C
FIFO_COUNTH = 0x72
FIFO_COUNTL = 0x73
FIFO_R_W = 0x74
WHO_AM_I = 0x75
def __init__(self, i2c, addr=0x68):
self.iic = i2c
self.addr = addr
self.iic.start()
self.iic.writeto(self.addr, bytearray([107, 0]))
self.iic.stop()
def _session(self):
self.iic.start()
yield
self.iic.stop()
def _read_from_mem(self, address, num_bytes):
self._session()
return self.iic.readfrom_mem(self.addr, address, num_bytes)
def _write_to_mem(self, address, values):
self._session()
self.iic.writeto_mem(self.addr, address, values)
def get_raw_values(self):
self.iic.start()
a = self.iic.readfrom_mem(self.addr, 0x3B, 14)
self.iic.stop()
return a
def get_ints(self):
b = self.get_raw_values()
c = []
for i in b:
c.append(i)
return c
def bytes_toint(self, firstbyte, secondbyte):
if not firstbyte & 0x80:
return firstbyte << 8 | secondbyte
return - (((firstbyte ^ 255) << 8) | (secondbyte ^ 255) + 1)
def get_values(self):
raw_ints = self.get_raw_values()
vals = {}
vals["AcX"] = self.bytes_toint(raw_ints[0], raw_ints[1])
vals["AcY"] = self.bytes_toint(raw_ints[2], raw_ints[3])
vals["AcZ"] = self.bytes_toint(raw_ints[4], raw_ints[5])
vals["Tmp"] = self.bytes_toint(raw_ints[6], raw_ints[7]) / 340.00 + 36.53
vals["GyX"] = self.bytes_toint(raw_ints[8], raw_ints[9])
vals["GyY"] = self.bytes_toint(raw_ints[10], raw_ints[11])
vals["GyZ"] = self.bytes_toint(raw_ints[12], raw_ints[13])
return vals # returned in range of Int16
# -32768 to 32767
def val_test(self): # ONLY FOR TESTING! Also, fast reading sometimes crashes IIC
from time import sleep
while 1:
print(self.get_values())
sleep(0.05)
| [
"carlos.barboza@gmail.com"
] | carlos.barboza@gmail.com |
8a5602d811f96072614c0fb9de6a9ce93ee0dc75 | 53095b893d0a80614aa43ab5a50b07d929fa4cf8 | /HRC_archive/LETG/extract_pha_file.py | 1cc120ec5884157a1a500717ef8a97b01372c3bf | [] | no_license | chandra-mta/HRC | e4cd297373e5e8f6dfb0e014275f5cee82b326b6 | aafb1f552394efd3367635a968f8df1e26b6f1c8 | refs/heads/main | 2021-10-23T23:06:45.964300 | 2021-10-21T10:10:09 | 2021-10-21T10:10:09 | 154,149,839 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 10,854 | py | #!/usr/bin/env /data/mta/Script/Python3.8/envs/ska3-shiny/bin/python
#########################################################################################
# #
# extract_pha_file.py: create a pha2 file and a tg directory for LETG observation #
# #
# author: t. isobe (tisobe@cfa.harvard.edu) #
# #
# last update: Apr 16, 2021 #
# #
#########################################################################################
import sys
import os
import string
import re
import math
import time
import astropy.io.fits as pyfits
#
#--- from ska
#
from Ska.Shell import getenv, bash
#
#--- set ciao environment
#
ciaoenv = getenv('source /soft/ciao/bin/ciao.csh -o', shell='tcsh')
#
#--- reading directory list
#
path = '/data/aschrc6/wilton/isobe/Project9/Scripts/LETG/house_keeping/dir_list'
with open(path, 'r') as f:
data = [line.strip() for line in f.readlines()]
for ent in data:
atemp = re.split(':', ent)
var = atemp[1].strip()
line = atemp[0].strip()
exec("%s = %s" %(var, line))
#
#--- append path to a private folders
#
sys.path.append(mta_dir)
sys.path.append(bin_dir)
import mta_common_functions as mcf
#
#--- temp writing file name
#
rtail = int(time.time())
zspace = '/tmp/zspace' + str(rtail)
#
#--- a couple of other settings
#
acmd = '/usr/bin/env PERL5LIB="" ;'
ilist = ['i', 's']
#------------------------------------------------------------------------------------
#-- extract_pha_file: create a pha2 file and a tg directory for LETG observation --
#------------------------------------------------------------------------------------
def extract_pha_file():
"""
create a pha2 file and a tg directory for LETG observation
input: none, but read from /data/hrc/<inst>
output: /data/hrc/<inst>/<obsid>/repro/*pha2.fits
/data/hrc/<inst>/<obsid>/repro/tg/*
"""
#
#--- get obsid list; d_list = [hrci_list, hrcs_list]
#
d_list = make_process_list()
for k in range(0, 2):
dlen = len(d_list[k])
for j in range(0, dlen):
obsid = d_list[k][j]
if j > 0:
cmd = 'rm -rf ./' + d_list[k][j-1]
os.system(cmd)
print("OBSID: " + str(obsid))
#
#--- original data directory name
#
d_dir = '/data/hrc/' + ilist[k] + '/' + str(obsid)
#
#--- copy the data to the local directory
#
cmd = 'cp -r ' + d_dir + ' ./' + str(obsid)
os.system(cmd)
#
#--- remove the files with "new" in the file names
#
cmd = 'rm -f ' + str(obsid) + '/secondary/*new*'
os.system(cmd)
#
#--- extract_pha_file chandra_repro
#
try:
pcmd = 'chandra_repro indir=' + str(obsid) + ' outdir='
pcmd = pcmd + str(obsid) + '/new cleanup=no'
cmd = acmd + pcmd
bash(cmd, env=ciaoenv)
except:
#
#--- if failed, keep the obsid in the record
#
ofile = house_keeping + 'no_repro_list'
with open(ofile, 'a') as fo:
eline = mcf.add_leading_zero(obsid, 5) + '\n'
fo.write(eline)
continue
#
#--- move pha2 file and tg directroy to the repro directory
#
outdir = d_dir + '/repro/'
cmd = 'mkdir -p ' + outdir #--- just in a case analysis dir does not exist
os.system(cmd)
cmd = 'mv -f ' + str(obsid) + '/new/*_pha2.fits* ' + outdir + '/.'
os.system(cmd)
cmd = 'mv -f ' + str(obsid) + '/new/tg ' + outdir + '/.'
os.system(cmd)
#
#--- change permission etc
#
cmd = 'chmod -R 775 ' + outdir
os.system(cmd)
cmd = 'chgrp -R hat ' + outdir
os.system(cmd)
#
#--- remove the copied data
#
cmd = 'rm -rf ./' + str(obsid)
os.system(cmd)
#
#--- fix naming to 5 digit obsid
#
correct_naming(obsid, ilist[k])
#
#--- send email
#
# line = 'HRC pha process finished\n'
# with open(zspace, 'w') as fo:
# fo.write(line)
#
# cmd = 'cat ' + zspace + '|mailx -s "Subject: HRC PHA finished" tisobe@cfa.harvard.edu'
# os.system(cmd)
# cmd = 'rm -rf ' + zspace
# os.system(cmd)
#------------------------------------------------------------------------------------
#-- make_process_list: create a list of unprocessed obsid lists --
#------------------------------------------------------------------------------------
def make_process_list():
"""
create a list of unprocessed obsid lists
input: none
output: a list of lists of [<hrc_i obsids>, <hrc_s obsids>]
"""
#
#--- create a dict: obsid <---> grating
#
[obs_list, dict_inst, dict_grat] = make_inst_dict()
#
#--- read failed repro obsid list
#
ifile = house_keeping + 'no_repro_list'
out = mcf.read_data_file(ifile)
rfailed = []
for ent in out:
rfailed.append(ent)
save = []
for inst in ['i', 's']:
hdir = '/data/hrc/' + inst + '/'
olist = []
#
#--- choose data with evt1 exists in the directory
#
cmd = 'ls -d ' + hdir + '*/secondary/*evt1.fits* > ' + zspace + ' 2>/dev/null'
os.system(cmd)
out = mcf.read_data_file(zspace, remove=1)
for ent in out:
atemp = re.split('\/', ent)
obsid = atemp[-3]
#
#--- check whether this obsid was previously checked, but failed to get the data
#
test = mcf.add_leading_zero(obsid, 5)
if test in rfailed:
continue
#
#--- check whether the pha2 file already exists
#
cmd = 'ls ' + hdir + obsid + '/repro/* > ' + zspace + ' 2>/dev/null'
os.system(cmd)
with open(zspace, 'r') as f:
ochk = f.read()
cmd = 'rm -rf ' + zspace
os.system(cmd)
mc = re.search('pha2', ochk)
#
#--- check whether it is an grating observation
#
if mc is None:
try:
iobsid = str(int(float(obsid)))
grat = dict_grat[iobsid]
except:
grat = 'NONE'
#
#--- special treatment for 6**** level calib observations
#
if obsid[0] == '6':
try:
grat = check_grating_from_header(inst, obsid)
except:
grat = 'NONE'
if grat == 'LETG':
olist.append(obsid)
save.append(olist)
return save
#------------------------------------------------------------------------------------
#-- make_inst_dict: create obsid <---> inst, obsid <---> grating dictionaries -
#------------------------------------------------------------------------------------
def make_inst_dict():
"""
create obsid <---> inst, obsid <---> grating dictionaries
input: none, but read from /data/mta4/obs_ss/sot_ocat.out
output: a list of <obsid list>, <dict of instruments>, <dict of grating>
note: only letg is taken as grating. hetg is ignored
"""
ifile = '/data/mta4/obs_ss/sot_ocat.out'
data = mcf.read_data_file(ifile)
obs_list = []
dict_inst = {}
dict_grat = {}
for ent in data:
mc1 = re.search('HRC', ent)
if mc1 is None:
continue
mc2 = re.search('archived', ent)
mc3 = re.search('observed', ent)
if (mc2 is None) and (mc3 is None):
continue
mc4 = re.search('LETG', ent)
if mc4 is not None:
grat = 'LETG'
else:
grat = 'NONE'
atemp = re.split('\^', ent)
obsid = atemp[1].strip()
obsid = str(int(float(obsid)))
inst = atemp[12].strip()
if inst in ['HRC-I', 'HRC-S']:
obs_list.append(obsid)
dict_inst[obsid] = inst
dict_grat[obsid] = grat
return [obs_list, dict_inst, dict_grat]
#------------------------------------------------------------------------------------
#-- check_grating_from_header: checking grating from a header of the evt1 file of obsid
#------------------------------------------------------------------------------------
def check_grating_from_header(hrc, obsid):
"""
checking grating from a header of the evt1 file of obsid
input: hrc --- either i or s
obsid --- obsid
output: grat --- gating, such as LETG, HETG, or NONE
"""
cmd = ' ls /data/hrc/' + hrc + '/' + obsid + '/secondary/*evt1.fits* > ' + zspace + ' 2>/dev/null'
os.system(cmd)
data = mcf.read_data_file(zspace, remove=1)
try:
fits = data[0].strip()
except:
return 'NONE'
flist = pyfits.open(fits)
try:
grat = flist[1].header['GRATING']
except:
grat = 'NONE'
flist.close()
return grat
#------------------------------------------------------------------------------------------------
#-- correct_naming: check repro directory and correct wrongly named fits and par file
#------------------------------------------------------------------------------------------------
def correct_naming(obsid, inst):
"""
check repro directory and correct wrongly named fits and par file
input: obsid --- obsid
inst --- instrument. either "i" or "s"
"""
cobsid = str(int(float(obsid)))
if len(cobsid) == 5:
return
lobsid = mcf.add_leading_zero(obsid, 5)
cmd = 'ls /data/hrc/' + inst + '/' + lobsid + '/repro/hrcf* >' + zspace
os.system(cmd)
data = mcf.read_data_file(zspace, remove=1)
for ent in data:
atemp = re.split('\/', ent)
fname = atemp[-1]
mc = re.search(lobsid, fname)
if mc is not None:
continue
else:
atemp = re.split('hrcf', fname)
btemp = re.split('_', atemp[1])
sobs = btemp[0]
new = fname.replace(sobs, lobsid)
full = '/data/hrc/' + inst + '/' + lobsid + '/' + sdir + '/' + new
cmd = 'mv ' + ent + ' ' + full
os.system(cmd)
#
#--- compress fits files
#
cmd = 'gzip /data/hrc/' + inst + '/' + lobsid + '/repro/*fits'
os.system(cmd)
#------------------------------------------------------------------------------------
if __name__ == "__main__":
extract_pha_file()
| [
"bspitzbart@cfa.harvard.edu"
] | bspitzbart@cfa.harvard.edu |
27bb5b5457ce8249495d9fcc5263dd01e827aed6 | e7d2c2c7fbcffc3b4e8976f01b354f794fc3b71d | /bmga/utils/formatting.py | 6ab11a6dfcbe02e87b651ba6371989e3ca2c1403 | [
"BSD-3-Clause"
] | permissive | vituocgia/boxme-api | 41da50fcec12089e59a29786b3bcff6c9b169d99 | 10c8054a223f124a85e70669d17313e3a2991226 | refs/heads/master | 2020-03-08T19:18:21.829490 | 2018-04-06T09:11:30 | 2018-04-06T09:11:30 | 128,347,542 | 0 | 0 | null | 2018-04-27T04:43:38 | 2018-04-06T05:24:05 | Python | UTF-8 | Python | false | false | 822 | py | from __future__ import unicode_literals
from django.utils import dateformat
from bmga.utils.timezone import make_naive, aware_datetime
from dateutil.parser import parse as mk_datetime # flake8: noqa
def format_datetime(dt):
"""
RFC 2822 datetime formatter
"""
return dateformat.format(make_naive(dt), 'r')
def format_date(d):
"""
RFC 2822 date formatter
"""
# workaround because Django's dateformat utility requires a datetime
# object (not just date)
dt = aware_datetime(d.year, d.month, d.day, 0, 0, 0)
return dateformat.format(dt, 'j M Y')
def format_time(t):
"""
RFC 2822 time formatter
"""
# again, workaround dateformat input requirement
dt = aware_datetime(2000, 1, 1, t.hour, t.minute, t.second)
return dateformat.format(dt, 'H:i:s O')
| [
"dotiendiep@gmail.com"
] | dotiendiep@gmail.com |
8a7e729fa2cc60c00b630bcd1e18b05703782c80 | 88d33f643f0116f0557a60d409183f66f528cb53 | /blogs/migrations/0043_auto_20190603_2054.py | ea9de7e0b4c9625696d3006dd57c7d5843a0b532 | [
"MIT"
] | permissive | vikrantgoutam17/my_blogs | 13e8e531a35173fbb22c07456b72444ba8391ee3 | 64b48eb8c466f46ac21e7ee137bbe8033f6a28ab | refs/heads/master | 2022-04-15T17:28:48.982593 | 2020-04-11T07:46:09 | 2020-04-11T07:46:09 | 254,733,039 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,059 | py | # Generated by Django 2.2.1 on 2019-06-03 15:24
import datetime
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('blogs', '0042_auto_20190603_1452'),
]
operations = [
migrations.AlterField(
model_name='blog',
name='pub_date',
field=models.DateTimeField(default=datetime.datetime(2019, 6, 3, 20, 54, 4, 48767), verbose_name='date published'),
),
migrations.CreateModel(
name='likes',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('blog_id', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='blogs.blog')),
('user', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)),
],
),
]
| [
"vikrantgoutam17@gmail.com"
] | vikrantgoutam17@gmail.com |
668403015d3558f567d49b9ab11c48be8202d6a0 | 69aaa71852bde5b2a4d297b6814ca0bf607053c8 | /jarvis.py | 6d46a86f112682ac4f06fe08a9938d84af0c382b | [] | no_license | sckzy/JarvisAI | cf36bf827a06faeb05a2a26d4f3a21dfd0eaa44c | 147b62cb2002b099865c10b2be023e0147fd96a1 | refs/heads/master | 2023-01-22T09:34:33.070981 | 2020-12-03T04:39:40 | 2020-12-03T04:39:40 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,217 | py | import pyttsx3
import datetime
import speech_recognition as sr
import wikipedia
import webbrowser
import os
import smtplib
def wishMe():
hour = int(datetime.datetime.now().hour)
if hour >= 0 and hour < 12:
speak("Good Morning!")
elif hour >= 12 and hour < 18:
speak("Good Afternoon!")
else:
speak("Good Evening!")
print("I am Jarvis Sir. Please tell me how may I help you")
speak("I am Jarvis Sir. Please tell me how may I help you")
print(
"System Info: system-started\nConnection to Internet: Connection Succeeded\nConnecting to Stelite No. 74: Done\n"
)
strTime = datetime.datetime.now().strftime("%H:%M:%S")
print("Status: time = ", strTime)
print("System Speed: 1Tb/s")
print("Internet Speed: 1Gbps")
engine = pyttsx3.init('sapi5')
voices = engine.getProperty('voices')
# print(voices[0].id)
engine.setProperty('voice', voices[0].id)
def speak(audio):
engine.say(audio)
engine.runAndWait()
def closecmd():
os.system('TASKKILL /F /IM chrome.exe')
def takeCommand():
#It takes microphone input from the user and returns string output
r = sr.Recognizer()
with sr.Microphone() as source:
print("Waiting for command Input...")
r.pause_threshold = 1
audio = r.listen(source)
try:
print("Identifying Command Input...")
query = r.recognize_google(audio, language='en-in')
print(f"Voice Input : {query}\n")
except Exception as e:
# print(e)
print("Sir please say that again I couldnt get It")
return "None"
return query
def sendEmail(to, content):
server = smtplib.SMTP('smtp.gmail.com', 587)
server.ehlo()
server.starttls()
server.login('your_email_addreess@gmail.com', 'password')
server.sendmail(dhangarjayashree51 @ gmail.com, to, content)
server.close()
def multiplication():
speak("PLease Enter the first Number")
print("PLease Enter the first Number")
a = takeCommand()
a = int(a)
speak("PLease Enter the second Number")
print("PLease Enter the Second Number")
b = takeCommand()
b = int(b)
product = a * b
print("The prduct is ", product)
speak(f"The answer is {product}")
def add():
speak("PLease Enter the first Number")
print("PLease Enter the first Number")
a = takeCommand()
a = int(a)
speak("PLease Enter the second Number")
print("PLease Enter the Second Number")
b = takeCommand()
b = int(b)
sum = a + b
print("The prduct is ", sum)
speak(f"The answer is {sum}")
def minus():
speak("Please Enter first number")
print("Please Enter first number")
a = takeCommand()
a = int(a)
speak("Please Enter second number")
print("Please Enter second number")
diff = a - b
print("The difference of the numbers is ", diff)
speak(f"The differnce is {diff}")
def devide():
speak("Please Enter first number")
print("Please Enter first number")
a = takeCommand()
a = int(a)
speak("Please Enter second number")
print("Please Enter second number")
que = a / b
print("The difference of the numbers is ", que)
speak(f"The differnce is {que}")
if __name__ == "__main__":
wishMe()
while True:
query = takeCommand().lower()
#Logic for executing commands
if 'wikipedia' in query:
speak('Searching Wikipedia...')
query = query.replace("wikipedia", "")
results = wikipedia.summary(query, sentences=2)
speak("According to Wikipedia")
speak(results)
elif 'open youtube' in query:
webbrowser.open("youtube.com")
elif 'open google' in query:
webbrowser.open("google.com")
elif 'open facebook' in query:
webbrowser.open("facebook.com")
elif 'open my website' in query:
webbrowser.open("https://codewithdhruva.github.io/MyWebsite/")
elif 'play music' in query:
music_dir = 'C:\\Users\\Prakash Dhangar\\Desktop\\My Projects\\Jarvis AI\\music'
songs = os.listdir(music_dir)
print(songs)
os.startfile(os.path.join(music_dir, songs[0]))
elif 'close' in query:
speak("What should I close")
print("Waiting for the answer")
if 'command prompt' in query:
os.system('TASKKILL /F /IM cmd.exe')
elif 'browser' in query:
os.system('TASKKILL /F /IM chrome.exe')
elif 'Browser' in query:
os.system('TASKKILL /F /IM chrome.exe')
elif 'code' in query:
os.systen('TASKKILL /F /IM code.exe')
elif 'the time' in query:
strTime = datetime.datetime.now().strftime("%H:%M:%S")
speak(f"Sir the time is {strTime}")
elif 'open code' in query:
print("Opening Code")
speak("Opening Code")
codePath = "C:\\Users\\Prakash Dhangar\\AppData\\Local\\Programs\\Microsoft VS Code\\Code.exe"
os.startfile(codePath)
elif 'open idea' in query:
print("Opening IntelliJ Idea")
speak("Opening IntelliJ Idea")
ideaPath = "C:\\Program Files\\JetBrains\\IntelliJ IDEA Community Edition 2020.1.2\\bin\\idea64.exe"
os.startfile(ideaPath)
elif 'open studio' in query:
print("Opening Android Studio")
speak("Opening Android Studio")
studioPath = "C:\\Program Files\\Android\\Android Studio\\bin\\studio64.exe"
os.startfile(studioPath)
elif 'play upgrade' in query:
upgradePath = "C:\\Users\\Prakash Dhangar\\Desktop\\Programms\\Upgrade.mkv"
speak("Playing Upgrade")
os.startfile(upgradePath)
elif 'send mail to prakash' in query:
try:
speak("What should I say?")
content = takeCommand()
to = "itworldcs@gmail.com"
sendEmail(to, content)
speak("Email has been sent!")
except Exception as e:
print(e)
speak("Sorry I couldnt send this Email")
elif 'thank you' in query:
speak(
"At your service I am always there. I am glad that I could help you."
)
elif 'open teams' in query:
teamsPath = "C:\\Users\\Prakash Dhangar\\AppData\\Roaming\\Microsoft\\Windows\\Start Menu\\Programs\\Microsoft Teams.ink"
speak("Starting Teams")
os.startfile(teamsPath)
elif 'play old music' in query:
webbrowser.open("https://youtu.be/kJIidWqWjUs?t=115")
elif 'multiply' in query:
multiplication()
elif 'sum of' in query:
add()
elif 'devide' in query:
devide()
elif 'minus' in query:
minus()
elif 'sleep' in query:
break
| [
"noreply@github.com"
] | noreply@github.com |
de74d6e820b635b6bbf1e43e970cb33fe5450482 | 0bfb74494054dc9e0774618687b5d234e6db6a92 | /ch12/gen_throw.py | fa2bf2ecf95d8d411c7a56100871ebf2f7390a2c | [] | no_license | FSSlc/AdvancePython | dddee32524cf1071a36eb4316efc2b53185645e3 | 61ed98f951fc877cc01e43a50488fc91b775c0f0 | refs/heads/master | 2022-11-06T10:53:07.812770 | 2020-06-25T23:43:24 | 2020-06-25T23:43:24 | 273,885,468 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 347 | py | #!/usr/bin/env python
# coding: utf-8
def gen_func():
try:
yield "user"
except Exception as e:
pass
yield 2
yield 3
return 'user'
if __name__ == '__main__':
gen = gen_func()
print(next(gen))
gen.throw(Exception, "download error")
print(next(gen))
gen.throw(Exception, "download error")
| [
"fsslc235@gmail.com"
] | fsslc235@gmail.com |
89b88fec9ac2985d22a7fca2da8155088f1e5567 | 219389652f57204b5d942ae8cdb7647650cc495f | /src/metricas.py | a92b9e13999190a65cc88303917aeaa402cc4b70 | [] | no_license | vsantsal/DesafioFolhaSistemaEstelarPython | dcc5bfb519ab1af3d022c642320511ef8a143115 | 19fa1e2102ec2348b2d5b7319efb2fd4efb66ddc | refs/heads/master | 2023-07-09T12:56:00.349140 | 2021-08-19T00:04:47 | 2021-08-19T00:04:47 | 267,566,330 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,661 | py | #!/usr/bin/env python3.9
from functools import reduce
from operator import mul
from typing import (
Any,
Dict,
List,
Protocol,
TypeVar,
)
"""
Classe responsável em que se a regra de negócio RN1 para descobrir o destino
da espaçonave está definida
RN1:O local de chegada pode ser conhecido sabendo que as vogais do nome da estrela
são atribuídas à uma sequência Fibonacci que começa em 1 e termina em 8,
onde A = 1, E = 2, I = 3, O = 5 e U = 8.
Se a multiplicação das vogais der o mesmo número
que a quantidade de engenheiros, a estrela de destino será conhecida.
"""
class Comparavel(Protocol):
def __eq__(self, other: Any) -> bool: ...
CT = TypeVar('CT', bound=Comparavel)
class Balanca:
@staticmethod
def compara(valor: CT, peso: CT):
# função que compara duas variáveis e retorna o booleano da comparação
return valor == peso
class Calculadora:
_MAPA_DE_VALORES: Dict[str, int] = {
'a': 1,
'e': 2,
'i': 3,
'o': 5,
'u': 8
}
def calcula(self, nome):
# converte a string fornecida em lista de inteiros
lista_a_multiplicar = self._converte_nome_em_lista_de_inteiros(nome)
# multiplica os elementos da lista de inteiros
resultado = reduce(mul, lista_a_multiplicar)
return resultado
def _converte_nome_em_lista_de_inteiros(self, nome: str) -> List[int]:
# inicializa lista para armazenar inteiros
resultado = [self._MAPA_DE_VALORES.get(letra.lower(),
1)
for letra in nome]
return resultado
| [
"vsantsal@gmail.com"
] | vsantsal@gmail.com |
a9f7a5d7afa2ce9694901a23dc8266394a8b4b54 | 1c6283303ceb883add8de4ee07c5ffcfc2e93fab | /Jinja2/lib/python3.7/site-packages/ixnetwork_restpy/testplatform/sessions/ixnetwork/vport/protocolstack/iptvrange_b15e81483d5be2cf30e042843e79b969.py | 1574bc8306f8407f267c617e318f6311dba42fcd | [] | no_license | pdobrinskiy/devcore | 0f5b3dfc2f3bf1e44abd716f008a01c443e14f18 | 580c7df6f5db8c118990cf01bc2b986285b9718b | refs/heads/main | 2023-07-29T20:28:49.035475 | 2021-09-14T10:02:16 | 2021-09-14T10:02:16 | 405,919,390 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 37,452 | py | # MIT LICENSE
#
# Copyright 1997 - 2020 by IXIA Keysight
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"),
# to deal in the Software without restriction, including without limitation
# the rights to use, copy, modify, merge, publish, distribute, sublicense,
# and/or sell copies of the Software, and to permit persons to whom the
# Software is furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in
# all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
# THE SOFTWARE.
from ixnetwork_restpy.base import Base
from ixnetwork_restpy.files import Files
from typing import List, Any, Union
class IptvRange(Base):
"""
The IptvRange class encapsulates a list of iptvRange resources that are managed by the user.
A list of resources can be retrieved from the server using the IptvRange.find() method.
The list can be managed by using the IptvRange.add() and IptvRange.remove() methods.
"""
__slots__ = ()
_SDM_NAME = 'iptvRange'
_SDM_ATT_MAP = {
'Enabled': 'enabled',
'GeneralQueryResponseMode': 'generalQueryResponseMode',
'ImmediateResponse': 'immediateResponse',
'InterStbStartDelay': 'interStbStartDelay',
'JoinLatencyThreshold': 'joinLatencyThreshold',
'JoinLeaveMultiplier': 'joinLeaveMultiplier',
'LeaveLatencyThreshold': 'leaveLatencyThreshold',
'LogFailureTimestamps': 'logFailureTimestamps',
'Name': 'name',
'ObjectId': 'objectId',
'ReportFrequency': 'reportFrequency',
'RouterAlert': 'routerAlert',
'SpecificQueryResponseMode': 'specificQueryResponseMode',
'StbLeaveJoinDelay': 'stbLeaveJoinDelay',
'UnsolicitedResponseMode': 'unsolicitedResponseMode',
'Version': 'version',
'ViewingProfile': 'viewingProfile',
}
_SDM_ENUM_MAP = {
}
def __init__(self, parent, list_op=False):
super(IptvRange, self).__init__(parent, list_op)
@property
def IptvChannels(self):
"""
Returns
-------
- obj(ixnetwork_restpy.testplatform.sessions.ixnetwork.vport.protocolstack.iptvchannels_7305b62e9ac2aa9f13637cc3a90a716f.IptvChannels): An instance of the IptvChannels class
Raises
------
- ServerError: The server has encountered an uncategorized error condition
"""
from ixnetwork_restpy.testplatform.sessions.ixnetwork.vport.protocolstack.iptvchannels_7305b62e9ac2aa9f13637cc3a90a716f import IptvChannels
if self._properties.get('IptvChannels', None) is not None:
return self._properties.get('IptvChannels')
else:
return IptvChannels(self)
@property
def Enabled(self):
# type: () -> bool
"""
Returns
-------
- bool: Disabled ranges won't be configured nor validated.
"""
return self._get_attribute(self._SDM_ATT_MAP['Enabled'])
@Enabled.setter
def Enabled(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['Enabled'], value)
@property
def GeneralQueryResponseMode(self):
# type: () -> bool
"""DEPRECATED
Returns
-------
- bool: If selected, responds to General Query messages.
"""
return self._get_attribute(self._SDM_ATT_MAP['GeneralQueryResponseMode'])
@GeneralQueryResponseMode.setter
def GeneralQueryResponseMode(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['GeneralQueryResponseMode'], value)
@property
def ImmediateResponse(self):
# type: () -> bool
"""DEPRECATED
Returns
-------
- bool: If selected, it will ignore the value specified in the Maximum Response Delay in the Membership Query message, assume that the Delay is always = 0 seconds and immediately respond to the Query by sending a Report.
"""
return self._get_attribute(self._SDM_ATT_MAP['ImmediateResponse'])
@ImmediateResponse.setter
def ImmediateResponse(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['ImmediateResponse'], value)
@property
def InterStbStartDelay(self):
# type: () -> int
"""
Returns
-------
- number: Time in milliseconds between Join messages from clients within the same range.
"""
return self._get_attribute(self._SDM_ATT_MAP['InterStbStartDelay'])
@InterStbStartDelay.setter
def InterStbStartDelay(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['InterStbStartDelay'], value)
@property
def JoinLatencyThreshold(self):
# type: () -> int
"""
Returns
-------
- number: The maximum time that is allowed for a multicast stream to arrive for channel for which a Join has been sent.
"""
return self._get_attribute(self._SDM_ATT_MAP['JoinLatencyThreshold'])
@JoinLatencyThreshold.setter
def JoinLatencyThreshold(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['JoinLatencyThreshold'], value)
@property
def JoinLeaveMultiplier(self):
# type: () -> int
"""DEPRECATED
Returns
-------
- number: The number of times a host sends every Join or Leave message.
"""
return self._get_attribute(self._SDM_ATT_MAP['JoinLeaveMultiplier'])
@JoinLeaveMultiplier.setter
def JoinLeaveMultiplier(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['JoinLeaveMultiplier'], value)
@property
def LeaveLatencyThreshold(self):
# type: () -> int
"""
Returns
-------
- number: The maximum time allowed for a multicast stream to stop for a channel for which a Leave has been sent.
"""
return self._get_attribute(self._SDM_ATT_MAP['LeaveLatencyThreshold'])
@LeaveLatencyThreshold.setter
def LeaveLatencyThreshold(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['LeaveLatencyThreshold'], value)
@property
def LogFailureTimestamps(self):
# type: () -> bool
"""
Returns
-------
- bool: If enabled, the timestamps for Join and Leave failures are saved to a log file.
"""
return self._get_attribute(self._SDM_ATT_MAP['LogFailureTimestamps'])
@LogFailureTimestamps.setter
def LogFailureTimestamps(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['LogFailureTimestamps'], value)
@property
def Name(self):
# type: () -> str
"""
Returns
-------
- str: Name of range
"""
return self._get_attribute(self._SDM_ATT_MAP['Name'])
@Name.setter
def Name(self, value):
# type: (str) -> None
self._set_attribute(self._SDM_ATT_MAP['Name'], value)
@property
def ObjectId(self):
# type: () -> str
"""
Returns
-------
- str: Unique identifier for this object
"""
return self._get_attribute(self._SDM_ATT_MAP['ObjectId'])
@property
def ReportFrequency(self):
# type: () -> int
"""DEPRECATED
Returns
-------
- number: When Send Unsolicited Response is enabled, specifies the frequency, in seconds, with which unsolicited messages are generated.
"""
return self._get_attribute(self._SDM_ATT_MAP['ReportFrequency'])
@ReportFrequency.setter
def ReportFrequency(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['ReportFrequency'], value)
@property
def RouterAlert(self):
# type: () -> bool
"""DEPRECATED
Returns
-------
- bool: If selected, sets the Send Router Alert bit in the IP header.
"""
return self._get_attribute(self._SDM_ATT_MAP['RouterAlert'])
@RouterAlert.setter
def RouterAlert(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['RouterAlert'], value)
@property
def SpecificQueryResponseMode(self):
# type: () -> bool
"""DEPRECATED
Returns
-------
- bool: If selected, responds to Group-Specific Query messages.
"""
return self._get_attribute(self._SDM_ATT_MAP['SpecificQueryResponseMode'])
@SpecificQueryResponseMode.setter
def SpecificQueryResponseMode(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['SpecificQueryResponseMode'], value)
@property
def StbLeaveJoinDelay(self):
# type: () -> int
"""
Returns
-------
- number: Time in milliseconds between sending a Leave for the current channel and Join for the next channel.
"""
return self._get_attribute(self._SDM_ATT_MAP['StbLeaveJoinDelay'])
@StbLeaveJoinDelay.setter
def StbLeaveJoinDelay(self, value):
# type: (int) -> None
self._set_attribute(self._SDM_ATT_MAP['StbLeaveJoinDelay'], value)
@property
def UnsolicitedResponseMode(self):
# type: () -> bool
"""DEPRECATED
Returns
-------
- bool: If selected, causes the emulated IGMP host to automatically send full membership messages at regular intervals, without waiting for a query message.
"""
return self._get_attribute(self._SDM_ATT_MAP['UnsolicitedResponseMode'])
@UnsolicitedResponseMode.setter
def UnsolicitedResponseMode(self, value):
# type: (bool) -> None
self._set_attribute(self._SDM_ATT_MAP['UnsolicitedResponseMode'], value)
@property
def Version(self):
# type: () -> str
"""DEPRECATED
Returns
-------
- str: IGMP/MLD protocol version.
"""
return self._get_attribute(self._SDM_ATT_MAP['Version'])
@Version.setter
def Version(self, value):
# type: (str) -> None
self._set_attribute(self._SDM_ATT_MAP['Version'], value)
@property
def ViewingProfile(self):
# type: () -> str
"""
Returns
-------
- str(None | /api/v1/sessions/1/ixnetwork/globals/.../iptvProfile): Template describing the behavior of how clients view the lists of channels.
"""
return self._get_attribute(self._SDM_ATT_MAP['ViewingProfile'])
@ViewingProfile.setter
def ViewingProfile(self, value):
# type: (str) -> None
self._set_attribute(self._SDM_ATT_MAP['ViewingProfile'], value)
def update(self, Enabled=None, GeneralQueryResponseMode=None, ImmediateResponse=None, InterStbStartDelay=None, JoinLatencyThreshold=None, JoinLeaveMultiplier=None, LeaveLatencyThreshold=None, LogFailureTimestamps=None, Name=None, ReportFrequency=None, RouterAlert=None, SpecificQueryResponseMode=None, StbLeaveJoinDelay=None, UnsolicitedResponseMode=None, Version=None, ViewingProfile=None):
# type: (bool, bool, bool, int, int, int, int, bool, str, int, bool, bool, int, bool, str, str) -> IptvRange
"""Updates iptvRange resource on the server.
Args
----
- Enabled (bool): Disabled ranges won't be configured nor validated.
- GeneralQueryResponseMode (bool): If selected, responds to General Query messages.
- ImmediateResponse (bool): If selected, it will ignore the value specified in the Maximum Response Delay in the Membership Query message, assume that the Delay is always = 0 seconds and immediately respond to the Query by sending a Report.
- InterStbStartDelay (number): Time in milliseconds between Join messages from clients within the same range.
- JoinLatencyThreshold (number): The maximum time that is allowed for a multicast stream to arrive for channel for which a Join has been sent.
- JoinLeaveMultiplier (number): The number of times a host sends every Join or Leave message.
- LeaveLatencyThreshold (number): The maximum time allowed for a multicast stream to stop for a channel for which a Leave has been sent.
- LogFailureTimestamps (bool): If enabled, the timestamps for Join and Leave failures are saved to a log file.
- Name (str): Name of range
- ReportFrequency (number): When Send Unsolicited Response is enabled, specifies the frequency, in seconds, with which unsolicited messages are generated.
- RouterAlert (bool): If selected, sets the Send Router Alert bit in the IP header.
- SpecificQueryResponseMode (bool): If selected, responds to Group-Specific Query messages.
- StbLeaveJoinDelay (number): Time in milliseconds between sending a Leave for the current channel and Join for the next channel.
- UnsolicitedResponseMode (bool): If selected, causes the emulated IGMP host to automatically send full membership messages at regular intervals, without waiting for a query message.
- Version (str): IGMP/MLD protocol version.
- ViewingProfile (str(None | /api/v1/sessions/1/ixnetwork/globals/.../iptvProfile)): Template describing the behavior of how clients view the lists of channels.
Raises
------
- ServerError: The server has encountered an uncategorized error condition
"""
return self._update(self._map_locals(self._SDM_ATT_MAP, locals()))
def add(self, Enabled=None, GeneralQueryResponseMode=None, ImmediateResponse=None, InterStbStartDelay=None, JoinLatencyThreshold=None, JoinLeaveMultiplier=None, LeaveLatencyThreshold=None, LogFailureTimestamps=None, Name=None, ReportFrequency=None, RouterAlert=None, SpecificQueryResponseMode=None, StbLeaveJoinDelay=None, UnsolicitedResponseMode=None, Version=None, ViewingProfile=None):
# type: (bool, bool, bool, int, int, int, int, bool, str, int, bool, bool, int, bool, str, str) -> IptvRange
"""Adds a new iptvRange resource on the server and adds it to the container.
Args
----
- Enabled (bool): Disabled ranges won't be configured nor validated.
- GeneralQueryResponseMode (bool): If selected, responds to General Query messages.
- ImmediateResponse (bool): If selected, it will ignore the value specified in the Maximum Response Delay in the Membership Query message, assume that the Delay is always = 0 seconds and immediately respond to the Query by sending a Report.
- InterStbStartDelay (number): Time in milliseconds between Join messages from clients within the same range.
- JoinLatencyThreshold (number): The maximum time that is allowed for a multicast stream to arrive for channel for which a Join has been sent.
- JoinLeaveMultiplier (number): The number of times a host sends every Join or Leave message.
- LeaveLatencyThreshold (number): The maximum time allowed for a multicast stream to stop for a channel for which a Leave has been sent.
- LogFailureTimestamps (bool): If enabled, the timestamps for Join and Leave failures are saved to a log file.
- Name (str): Name of range
- ReportFrequency (number): When Send Unsolicited Response is enabled, specifies the frequency, in seconds, with which unsolicited messages are generated.
- RouterAlert (bool): If selected, sets the Send Router Alert bit in the IP header.
- SpecificQueryResponseMode (bool): If selected, responds to Group-Specific Query messages.
- StbLeaveJoinDelay (number): Time in milliseconds between sending a Leave for the current channel and Join for the next channel.
- UnsolicitedResponseMode (bool): If selected, causes the emulated IGMP host to automatically send full membership messages at regular intervals, without waiting for a query message.
- Version (str): IGMP/MLD protocol version.
- ViewingProfile (str(None | /api/v1/sessions/1/ixnetwork/globals/.../iptvProfile)): Template describing the behavior of how clients view the lists of channels.
Returns
-------
- self: This instance with all currently retrieved iptvRange resources using find and the newly added iptvRange resources available through an iterator or index
Raises
------
- ServerError: The server has encountered an uncategorized error condition
"""
return self._create(self._map_locals(self._SDM_ATT_MAP, locals()))
def remove(self):
"""Deletes all the contained iptvRange resources in this instance from the server.
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
self._delete()
def find(self, Enabled=None, GeneralQueryResponseMode=None, ImmediateResponse=None, InterStbStartDelay=None, JoinLatencyThreshold=None, JoinLeaveMultiplier=None, LeaveLatencyThreshold=None, LogFailureTimestamps=None, Name=None, ObjectId=None, ReportFrequency=None, RouterAlert=None, SpecificQueryResponseMode=None, StbLeaveJoinDelay=None, UnsolicitedResponseMode=None, Version=None, ViewingProfile=None):
# type: (bool, bool, bool, int, int, int, int, bool, str, str, int, bool, bool, int, bool, str, str) -> IptvRange
"""Finds and retrieves iptvRange resources from the server.
All named parameters are evaluated on the server using regex. The named parameters can be used to selectively retrieve iptvRange resources from the server.
To retrieve an exact match ensure the parameter value starts with ^ and ends with $
By default the find method takes no parameters and will retrieve all iptvRange resources from the server.
Args
----
- Enabled (bool): Disabled ranges won't be configured nor validated.
- GeneralQueryResponseMode (bool): If selected, responds to General Query messages.
- ImmediateResponse (bool): If selected, it will ignore the value specified in the Maximum Response Delay in the Membership Query message, assume that the Delay is always = 0 seconds and immediately respond to the Query by sending a Report.
- InterStbStartDelay (number): Time in milliseconds between Join messages from clients within the same range.
- JoinLatencyThreshold (number): The maximum time that is allowed for a multicast stream to arrive for channel for which a Join has been sent.
- JoinLeaveMultiplier (number): The number of times a host sends every Join or Leave message.
- LeaveLatencyThreshold (number): The maximum time allowed for a multicast stream to stop for a channel for which a Leave has been sent.
- LogFailureTimestamps (bool): If enabled, the timestamps for Join and Leave failures are saved to a log file.
- Name (str): Name of range
- ObjectId (str): Unique identifier for this object
- ReportFrequency (number): When Send Unsolicited Response is enabled, specifies the frequency, in seconds, with which unsolicited messages are generated.
- RouterAlert (bool): If selected, sets the Send Router Alert bit in the IP header.
- SpecificQueryResponseMode (bool): If selected, responds to Group-Specific Query messages.
- StbLeaveJoinDelay (number): Time in milliseconds between sending a Leave for the current channel and Join for the next channel.
- UnsolicitedResponseMode (bool): If selected, causes the emulated IGMP host to automatically send full membership messages at regular intervals, without waiting for a query message.
- Version (str): IGMP/MLD protocol version.
- ViewingProfile (str(None | /api/v1/sessions/1/ixnetwork/globals/.../iptvProfile)): Template describing the behavior of how clients view the lists of channels.
Returns
-------
- self: This instance with matching iptvRange resources retrieved from the server available through an iterator or index
Raises
------
- ServerError: The server has encountered an uncategorized error condition
"""
return self._select(self._map_locals(self._SDM_ATT_MAP, locals()))
def read(self, href):
"""Retrieves a single instance of iptvRange data from the server.
Args
----
- href (str): An href to the instance to be retrieved
Returns
-------
- self: This instance with the iptvRange resources from the server available through an iterator or index
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
return self._read(href)
def CustomProtocolStack(self, *args, **kwargs):
# type: (*Any, **Any) -> None
"""Executes the customProtocolStack operation on the server.
Create custom protocol stack under /vport/protocolStack
customProtocolStack(Arg2=list, Arg3=enum, async_operation=bool)
---------------------------------------------------------------
- Arg2 (list(str)): List of plugin types to be added in the new custom stack
- Arg3 (str(kAppend | kMerge | kOverwrite)): Append, merge or overwrite existing protocol stack
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
payload = { "Arg1": self }
for i in range(len(args)): payload['Arg%s' % (i + 2)] = args[i]
for item in kwargs.items(): payload[item[0]] = item[1]
return self._execute('customProtocolStack', payload=payload, response_object=None)
def DisableProtocolStack(self, *args, **kwargs):
# type: (*Any, **Any) -> Union[str, None]
"""Executes the disableProtocolStack operation on the server.
Disable a protocol under protocolStack using the class name
disableProtocolStack(Arg2=string, async_operation=bool)string
-------------------------------------------------------------
- Arg2 (str): Protocol class name to disable
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
- Returns str: Status of the exec
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
payload = { "Arg1": self.href }
for i in range(len(args)): payload['Arg%s' % (i + 2)] = args[i]
for item in kwargs.items(): payload[item[0]] = item[1]
return self._execute('disableProtocolStack', payload=payload, response_object=None)
def EnableProtocolStack(self, *args, **kwargs):
# type: (*Any, **Any) -> Union[str, None]
"""Executes the enableProtocolStack operation on the server.
Enable a protocol under protocolStack using the class name
enableProtocolStack(Arg2=string, async_operation=bool)string
------------------------------------------------------------
- Arg2 (str): Protocol class name to enable
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
- Returns str: Status of the exec
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
payload = { "Arg1": self.href }
for i in range(len(args)): payload['Arg%s' % (i + 2)] = args[i]
for item in kwargs.items(): payload[item[0]] = item[1]
return self._execute('enableProtocolStack', payload=payload, response_object=None)
def IptvStart(self, *args, **kwargs):
# type: (*Any, **Any) -> None
"""Executes the iptvStart operation on the server.
Start IPTV on selected plugins and ranges
The IxNetwork model allows for multiple method Signatures with the same name while python does not.
iptvStart(async_operation=bool)
-------------------------------
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
iptvStart(Arg2=enum, async_operation=bool)
------------------------------------------
- Arg2 (str(async | sync)): kArray[kObjref=/vport/protocolStack/atm/dhcpEndpoint/iptv,/vport/protocolStack/atm/dhcpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/dhcpEndpoint/iptv,/vport/protocolStack/atm/emulatedRouter/dhcpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/iptv,/vport/protocolStack/atm/emulatedRouter/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ipEndpoint/iptv,/vport/protocolStack/atm/emulatedRouter/ipEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/iptv,/vport/protocolStack/atm/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/atm/ipEndpoint/iptv,/vport/protocolStack/atm/ipEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/dhcpoPppClientEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/dhcpoPppServerEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/iptv,/vport/protocolStack/atm/pppoxEndpoint/iptv,/vport/protocolStack/atm/pppoxEndpoint/range/iptvRange,/vport/protocolStack/ethernet/dhcpEndpoint/iptv,/vport/protocolStack/ethernet/dhcpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/dhcpEndpoint/iptv,/vport/protocolStack/ethernet/emulatedRouter/dhcpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/iptv,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ipEndpoint/iptv,/vport/protocolStack/ethernet/emulatedRouter/ipEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/iptv,/vport/protocolStack/ethernet/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ipEndpoint/iptv,/vport/protocolStack/ethernet/ipEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/dhcpoPppClientEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/dhcpoPppServerEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/iptv,/vport/protocolStack/ethernet/pppoxEndpoint/iptv,/vport/protocolStack/ethernet/pppoxEndpoint/range/iptvRange]
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
payload = { "Arg1": self }
for i in range(len(args)): payload['Arg%s' % (i + 2)] = args[i]
for item in kwargs.items(): payload[item[0]] = item[1]
return self._execute('iptvStart', payload=payload, response_object=None)
def IptvStop(self, *args, **kwargs):
# type: (*Any, **Any) -> None
"""Executes the iptvStop operation on the server.
Stop IPTV on selected plugins and ranges
The IxNetwork model allows for multiple method Signatures with the same name while python does not.
iptvStop(async_operation=bool)
------------------------------
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
iptvStop(Arg2=enum, async_operation=bool)
-----------------------------------------
- Arg2 (str(async | sync)): kArray[kObjref=/vport/protocolStack/atm/dhcpEndpoint/iptv,/vport/protocolStack/atm/dhcpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/dhcpEndpoint/iptv,/vport/protocolStack/atm/emulatedRouter/dhcpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/iptv,/vport/protocolStack/atm/emulatedRouter/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/atm/emulatedRouter/ipEndpoint/iptv,/vport/protocolStack/atm/emulatedRouter/ipEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/iptv,/vport/protocolStack/atm/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/atm/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/atm/ipEndpoint/iptv,/vport/protocolStack/atm/ipEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/dhcpoPppClientEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/dhcpoPppServerEndpoint/range/iptvRange,/vport/protocolStack/atm/pppox/iptv,/vport/protocolStack/atm/pppoxEndpoint/iptv,/vport/protocolStack/atm/pppoxEndpoint/range/iptvRange,/vport/protocolStack/ethernet/dhcpEndpoint/iptv,/vport/protocolStack/ethernet/dhcpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/dhcpEndpoint/iptv,/vport/protocolStack/ethernet/emulatedRouter/dhcpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/iptv,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/emulatedRouter/ipEndpoint/iptv,/vport/protocolStack/ethernet/emulatedRouter/ipEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpPcrfEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpPcrfS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpS5S8PgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpSgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/egtpUeS5S8SgwEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/iptv,/vport/protocolStack/ethernet/ip/l2tp/dhcpoLacEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/l2tp/dhcpoLnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/l2tpEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ip/smDnsEndpoint/range/iptvRange,/vport/protocolStack/ethernet/ipEndpoint/iptv,/vport/protocolStack/ethernet/ipEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/dhcpoPppClientEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/dhcpoPppServerEndpoint/range/iptvRange,/vport/protocolStack/ethernet/pppox/iptv,/vport/protocolStack/ethernet/pppoxEndpoint/iptv,/vport/protocolStack/ethernet/pppoxEndpoint/range/iptvRange]
- async_operation (bool=False): True to execute the operation asynchronously. Any subsequent rest api calls made through the Connection class will block until the operation is complete.
Raises
------
- NotFoundError: The requested resource does not exist on the server
- ServerError: The server has encountered an uncategorized error condition
"""
payload = { "Arg1": self }
for i in range(len(args)): payload['Arg%s' % (i + 2)] = args[i]
for item in kwargs.items(): payload[item[0]] = item[1]
return self._execute('iptvStop', payload=payload, response_object=None)
| [
"pdobrinskiy@yahoo.com"
] | pdobrinskiy@yahoo.com |
d57deec55818ba052668da21c0a7d21e7d37f27c | 3365209124a099d96b7397012865a544c60fbab3 | /PyPoll/PyPoll_IGR.py | 3f65ff292c2d4a187c26ced1618990ca64bd6531 | [] | no_license | igr1993/python-challenge | 317f651c154edc862907998e543e11ceaab6f07e | 857977490b622c6f7bf4edf0e0f0065c91149959 | refs/heads/main | 2023-03-07T06:04:02.207958 | 2021-02-21T01:41:07 | 2021-02-21T01:41:07 | 340,744,301 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,134 | py |
#lets go PyPoll!
import os
import csv
candidates = {}
election_csv = os.path.join('.','Resources','election_data.csv')
with open(election_csv, 'r') as csvfile:
csvreader = csv.reader(csvfile, delimiter=',')
header = next(csvreader)
for row in csvreader:
if row[2] in candidates.keys():
candidates[row[2]]+=1
else:
candidates[row[2]] = 1
total = candidates.values()
total_votes = sum(total)
list_candidates = candidates.keys()
votes_per = [f'{(x/total_votes)*100:.3f}%' for x in candidates.values()]
winner = list(candidates.keys())[list(candidates.values()).index(max(candidates.values()))]
winner
print("Election results")
print("--------------------------------")
print(f" Total votes: {int(total_votes)}")
print("---------------------------------")
i = 0
for candidate, vote in candidates.items():
print(f'{candidate}, {vote} , {votes_per[i]}')
i+=1
print("------------------------------")
print(f" Winner: {winner}")
print("------------------------------") | [
"igr@Ians-MacBook-Pro.local"
] | igr@Ians-MacBook-Pro.local |
09e6ba57f63d8ca9d88d39ff34881ab5d048bc96 | e823bc36af457f229f6879d6e6a3ef6247c129aa | /virtualenv/Lib/site-packages/pyasn1_modules/rfc2560.py | e41994abf64813e9eb5f70b1dd1b03519a320ed1 | [
"MIT"
] | permissive | William-An/DFB_Final | e772fa979c41f2f83a4bf657cde499456215fb3b | 49a9244c98116574676992ebecd1d9435e1d5b1e | refs/heads/master | 2022-11-07T15:47:36.189057 | 2017-07-22T01:01:37 | 2017-07-22T01:01:43 | 97,426,562 | 1 | 1 | MIT | 2022-10-15T02:45:57 | 2017-07-17T02:21:42 | Python | UTF-8 | Python | false | false | 8,251 | py | #
# This file is part of pyasn1-modules software.
#
# Copyright (c) 2005-2017, Ilya Etingof <etingof@gmail.com>
# License: http://pyasn1.sf.net/license.html
#
# OCSP request/response syntax
#
# Derived from a minimal OCSP library (RFC2560) code written by
# Bud P. Bruegger <bud@ancitel.it>
# Copyright: Ancitel, S.p.a, Rome, Italy
# License: BSD
#
#
# current limitations:
# * request and response works only for a single certificate
# * only some values are parsed out of the response
# * the request does't set a nonce nor signature
# * there is no signature validation of the response
# * dates are left as strings in GeneralizedTime format -- datetime.datetime
# would be nicer
#
from pyasn1.type import tag, namedtype, namedval, univ, useful
from pyasn1_modules import rfc2459
# Start of OCSP module definitions
# This should be in directory Authentication Framework (X.509) module
class CRLReason(univ.Enumerated):
namedValues = namedval.NamedValues(
('unspecified', 0),
('keyCompromise', 1),
('cACompromise', 2),
('affiliationChanged', 3),
('superseded', 4),
('cessationOfOperation', 5),
('certificateHold', 6),
('removeFromCRL', 8),
('privilegeWithdrawn', 9),
('aACompromise', 10)
)
# end of directory Authentication Framework (X.509) module
# This should be in PKIX Certificate Extensions module
class GeneralName(univ.OctetString):
pass
# end of PKIX Certificate Extensions module
id_kp_OCSPSigning = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 3, 9))
id_pkix_ocsp = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1))
id_pkix_ocsp_basic = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 1))
id_pkix_ocsp_nonce = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 2))
id_pkix_ocsp_crl = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 3))
id_pkix_ocsp_response = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 4))
id_pkix_ocsp_nocheck = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 5))
id_pkix_ocsp_archive_cutoff = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 6))
id_pkix_ocsp_service_locator = univ.ObjectIdentifier((1, 3, 6, 1, 5, 5, 7, 48, 1, 7))
class AcceptableResponses(univ.SequenceOf):
componentType = univ.ObjectIdentifier()
class ArchiveCutoff(useful.GeneralizedTime):
pass
class UnknownInfo(univ.Null):
pass
class RevokedInfo(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('revocationTime', useful.GeneralizedTime()),
namedtype.OptionalNamedType('revocationReason', CRLReason().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
class CertID(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('hashAlgorithm', rfc2459.AlgorithmIdentifier()),
namedtype.NamedType('issuerNameHash', univ.OctetString()),
namedtype.NamedType('issuerKeyHash', univ.OctetString()),
namedtype.NamedType('serialNumber', rfc2459.CertificateSerialNumber())
)
class CertStatus(univ.Choice):
componentType = namedtype.NamedTypes(
namedtype.NamedType('good',
univ.Null().subtype(implicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0))),
namedtype.NamedType('revoked',
RevokedInfo().subtype(implicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 1))),
namedtype.NamedType('unknown',
UnknownInfo().subtype(implicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 2)))
)
class SingleResponse(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('certID', CertID()),
namedtype.NamedType('certStatus', CertStatus()),
namedtype.NamedType('thisUpdate', useful.GeneralizedTime()),
namedtype.OptionalNamedType('nextUpdate', useful.GeneralizedTime().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0))),
namedtype.OptionalNamedType('singleExtensions', rfc2459.Extensions().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 1)))
)
class KeyHash(univ.OctetString):
pass
class ResponderID(univ.Choice):
componentType = namedtype.NamedTypes(
namedtype.NamedType('byName',
rfc2459.Name().subtype(implicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 1))),
namedtype.NamedType('byKey',
KeyHash().subtype(implicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 2)))
)
class Version(univ.Integer):
namedValues = namedval.NamedValues(('v1', 0))
class ResponseData(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.DefaultedNamedType('version', Version('v1').subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0))),
namedtype.NamedType('responderID', ResponderID()),
namedtype.NamedType('producedAt', useful.GeneralizedTime()),
namedtype.NamedType('responses', univ.SequenceOf(SingleResponse())),
namedtype.OptionalNamedType('responseExtensions', rfc2459.Extensions().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 1)))
)
class BasicOCSPResponse(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('tbsResponseData', ResponseData()),
namedtype.NamedType('signatureAlgorithm', rfc2459.AlgorithmIdentifier()),
namedtype.NamedType('signature', univ.BitString()),
namedtype.OptionalNamedType('certs', univ.SequenceOf(rfc2459.Certificate()).subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
class ResponseBytes(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('responseType', univ.ObjectIdentifier()),
namedtype.NamedType('response', univ.OctetString())
)
class OCSPResponseStatus(univ.Enumerated):
namedValues = namedval.NamedValues(
('successful', 0),
('malformedRequest', 1),
('internalError', 2),
('tryLater', 3),
('undefinedStatus', 4), # should never occur
('sigRequired', 5),
('unauthorized', 6)
)
class OCSPResponse(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('responseStatus', OCSPResponseStatus()),
namedtype.OptionalNamedType('responseBytes', ResponseBytes().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
class Request(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('reqCert', CertID()),
namedtype.OptionalNamedType('singleRequestExtensions', rfc2459.Extensions().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
class Signature(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('signatureAlgorithm', rfc2459.AlgorithmIdentifier()),
namedtype.NamedType('signature', univ.BitString()),
namedtype.OptionalNamedType('certs', univ.SequenceOf(rfc2459.Certificate()).subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
class TBSRequest(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.DefaultedNamedType('version', Version('v1').subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0))),
namedtype.OptionalNamedType('requestorName', GeneralName().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 1))),
namedtype.NamedType('requestList', univ.SequenceOf(Request())),
namedtype.OptionalNamedType('requestExtensions', rfc2459.Extensions().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 2)))
)
class OCSPRequest(univ.Sequence):
componentType = namedtype.NamedTypes(
namedtype.NamedType('tbsRequest', TBSRequest()),
namedtype.OptionalNamedType('optionalSignature', Signature().subtype(
explicitTag=tag.Tag(tag.tagClassContext, tag.tagFormatSimple, 0)))
)
| [
"China_Aisa@live.com"
] | China_Aisa@live.com |
3f1407aeca182aa219f12783beac7371b4dd6ad1 | 3c3df48ff6239c5387847200c25d67c8c1b59491 | /alien_invasion.py | e53683fdd6c8210596db0118bfe98acfcdf4cf9c | [] | no_license | sdkcouto/coronapython-game | cf8930721883f9ec7786d61959989bcdf2a381d0 | 5e11f35d6ae8035d640d4e768432074f7a80f427 | refs/heads/master | 2022-11-19T03:47:27.375952 | 2020-07-21T18:44:25 | 2020-07-21T18:44:25 | 278,443,053 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,208 | py | import pygame
from pygame.sprite import Group
from settings import Settings
from game_stats import GameStats
from scoreboard import Scoreboard
from button import Button
from ship import Ship
from alien import Alien
import game_functions as gf
def run_game():
pygame.init()
ai_settings = Settings()
screen = pygame.display.set_mode((ai_settings.screen_width, ai_settings.screen_height))
pygame.display.set_caption("Alien Invasion")
play_button = Button(ai_settings, screen, "Play")
stats = GameStats(ai_settings)
sb = Scoreboard(ai_settings, screen, stats)
ship = Ship(ai_settings, screen)
aliens = Group()
bullets = Group()
gf.create_fleet(ai_settings, screen, ship, aliens)
while True:
gf.check_events(ai_settings, screen, stats, sb, play_button, ship, aliens, bullets)
if stats.game_active:
ship.update()
bullets.update()
gf.update_bullets(ai_settings, screen, stats, sb, ship, aliens, bullets)
gf.update_aliens(ai_settings, screen, stats, sb, ship, aliens, bullets)
gf.update_screen(ai_settings, screen, stats, sb, ship, aliens, bullets, play_button)
run_game() | [
"matheuscouto.eng@gmail.com"
] | matheuscouto.eng@gmail.com |
b23b322eb6d1bfaf868d64b5b3e0f1fc530701a6 | d783f5e6b02d91990e30b7e9edb41b9725f540cf | /intro-cs/intro-to-programming/exercises/ex_06_02.py | f7ca5b3c1d5f30d22db7facaa3d2df4d7888af4c | [
"MIT"
] | permissive | solanyn/cs-courses | 1c4ac7ea146b816731f7016c988423a8a5c7d180 | db10f8d9ad9a4f00a661c58eed7d2423ff9871bc | refs/heads/main | 2023-07-05T16:51:49.384749 | 2021-08-06T04:01:38 | 2021-08-06T04:01:38 | 332,115,999 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 128 | py | """Exercise 2: Given that fruit is a string, what does fruit[:] mean?"""
# fruit[:] means all characters in the sequence fruit
| [
"andrewchen1520@gmail.com"
] | andrewchen1520@gmail.com |
17f9e6c0e56e8413c26b56d50634d33d99988d02 | 0dcb743d6409478d01706f48616df6a1c089c0f0 | /Project/code/loadbalancer-server.py | 097411dbfc2d4e9664c83cfd5f70331a2709c1ef | [] | no_license | marcotcr/systems | 063b2b78b99e9e0a7733760125a999106d7fc6ba | b35358cbc83daa7900f049bf544bd4df500b0749 | refs/heads/master | 2021-01-20T10:50:28.518802 | 2014-03-14T16:42:32 | 2014-03-14T16:42:32 | 15,811,702 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 2,224 | py | from __future__ import division
import sys
sys.path.append('gen-py/autoscale')
sys.path.append('gen-py/')
import LoadBalancer
from autoscale.ttypes import *
from autoscale import LoadBalancer
from thrift.transport import TSocket
from thrift.transport import TTransport
from thrift.protocol import TBinaryProtocol
from thrift.server import TServer
import collections
import socket
import time
import argparse
import numpy as np
import os
import random
import threading
from itertools import cycle
class LoadBalancerHandler:
def __init__(self):
self.node_times = {}
self.nodelist = []
self.n_cycle = cycle(range(0))
self.n = 0
self.n_threads = 10
pass
def Finish(self, node):
time.sleep(self.node_times[node])
self.n_running[node] -= 1
def GetNode(self):
# print 'GetNode'
node = self.nodelist[self.n_cycle.next()]
while self.n_running[node] >= self.n_threads:
node = self.nodelist[self.n_cycle.next()]
self.n_running[node] += 1
self.n += 1
t = threading.Thread(target=self.Finish, args=(node,))
t.start()
return node
def SetNodes(self, state):
print 'SetNodes', state
self.node_times = {}
self.previous_state = state
for node, quota in state.iteritems():
self.node_times[node] = float(quota)
self.nodelist = self.node_times.keys()
self.n_running = collections.defaultdict(lambda: 0)
self.n_cycle = cycle(range(len(self.nodelist)))
def NumRequests(self):
return self.n
def main():
parser = argparse.ArgumentParser(description='TODO')
parser.add_argument('-p', '--port', type=int, required=True, help="My port")
args = parser.parse_args()
port = args.port
handler = LoadBalancerHandler()
processor = LoadBalancer.Processor(handler)
transport = TSocket.TServerSocket(port=port)
tfactory = TTransport.TBufferedTransportFactory()
pfactory = TBinaryProtocol.TBinaryProtocolFactory()
server = TServer.TThreadPoolServer(processor, transport, tfactory, pfactory)
server.setNumThreads(10)
# server = TServer.TThreadedServer(processor, transport, tfactory, pfactory)
print "Starting python server..."
server.serve()
print "done!"
if __name__ == '__main__':
main()
| [
"marcotcr@gmail.com"
] | marcotcr@gmail.com |
c2f5d321adc23f137909c4b7b15ea48ad477e908 | d9874a4d69027dd401314828a9d19e686a62b31d | /stamps.py | 799b1f7b60be0575dc81351a41ef953ca79ab61a | [] | no_license | 09o/ProgrammingBasis | e9612b7ac59aadf6cf406dbbb7bb42c5d9502e42 | 6fdd6b69653cc6d89b6841883e6795191a742847 | refs/heads/master | 2020-04-03T13:03:30.455309 | 2018-11-14T08:44:54 | 2018-11-14T08:44:54 | 155,272,452 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,270 | py | # Define a procedure, stamps, which takes as its input a positive integer in
# pence and returns the number of 5p, 2p and 1p stamps (p is pence) required
# to make up that value. The return value should be a tuple of three numbers
# (that is, your return statement should be followed by the number of 5p,
# the number of 2p, and the nuber of 1p stamps).
#
# Your answer should use as few total stamps as possible by first using as
# many 5p stamps as possible, then 2 pence stamps and finally 1p stamps as
# needed to make up the total.
#
# (No fair for USians to just say use a "Forever" stamp and be done with it!)
#
def stamps(n):
i = j = k = 0
if n != 0:
while i*5 <= n:
i += 1
i -= 1
n -= i*5
while j*2 <= n:
j += 1
j -= 1
n -= j*2
if n == 1:
k = 1
else:
k = 0
else:
return i,j,k
return i,j,k
# FOR TEST
print(stamps(8))
#>>> (1, 1, 1) # one 5p stamp, one 2p stamp and one 1p stamp
print(stamps(5))
#>>> (1, 0, 0) # one 5p stamp, no 2p stamps and no 1p stamps
print(stamps(29))
#>>> (5, 2, 0) # five 5p stamps, two 2p stamps and no 1p stamps
print(stamps(0))
#>>> (0, 0, 0) # no 5p stamps, no 2p stamps and no 1p stamps
| [
"igsober@gmail.com"
] | igsober@gmail.com |
76d236d6b4d1a10f3a74eee09e3478109205972e | 830db802b4cce9d4b5a34c8351caf6573bf5f385 | /basic_functions.py | 2da3b0a3d1d10dea785c7a1b0ddb5a0be4d7addb | [] | no_license | 317002/Normal-Modes | 856e33dbe103c45ab17e9c0c417a7f8bbd64aed8 | dc599e1c4abc035bae8abdced600119ec42fc7df | refs/heads/master | 2020-08-31T22:59:41.116273 | 2019-11-12T01:46:38 | 2019-11-12T01:46:38 | 218,807,505 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,471 | py | # -*- coding: utf-8 -*-
"""
Created on Fri Nov 8 14:38:11 2019
@author: Nathan Richard
"""
import numpy as np
from math import floor
def gen_k(n):
index = []
ind = []
for i in range(0,n+2):
ind += [0]
for i in range(0,n+2):
index += [ind]
for i in range(0,len(index)):
for j in range(0,len(index)):
index[i][j] = n**2
index = np.array(index)
c = 0
for i in range(1,n+1):
for j in range(1,n+1):
index[i][j] = c
c += 1
K = []
for i in range(1,n+1):
for j in range(1,n+1):
m = []
for l in range(0,n**2+1):
m += [0]
m[index[i][j]] = -4
m[index[i+1][j]] = 1
m[index[i-1][j]] = 1
m[index[i][j+1]] = 1
m[index[i][j-1]] = 1
K += [m[:-1]]
K = np.array(K)
return(-K)
def gen_square_posistion_matrix(N):
if N%2 != 1:
#checking if the value is odd
raise Exception('N must be odd')
'''So the matrix can be centered around zero'''
if N == 1:
a = 1
return np.asarray([1]),np.asarray([1])
else:
a = floor(N/2)#gives me the exstent of the x vector
x = np.linspace(-a,a,N)
y = x.copy()
x = np.zeros((x.shape[0],1),dtype=x.dtype) + x
y = x.transpose()
return (x,y) | [
"ndr1005@wildcats.unh.edu"
] | ndr1005@wildcats.unh.edu |
29fb1dc505940b341e55979260d1600d60d2fd10 | 6fad4c35342a27bd9882b3a55e99da461df8f6c3 | /scripts/token_check.py | 81bcb49afee511a7d3b182e99de819cba6c8a4f2 | [
"MIT"
] | permissive | tonychew1986/erc20-token | 137bdb06dbbd275165cf6f9c6703b6927c9c4cb3 | dd206c2e2282f0fce57c4da05111210102bbfbde | refs/heads/main | 2023-03-28T01:12:14.142687 | 2021-03-28T08:54:10 | 2021-03-28T08:54:10 | 352,282,111 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 405 | py | #!/usr/bin/python3
from brownie import Token, accounts
#etherscan api
#Q9JWR2ZC58AGT4MTM7CEFI8UD8IH3VQ2WS
def main():
# user_account = accounts.load('xxx_shared_1')
# user_account = accounts.load('xxx_shared_mainnet')
token_contract = Token.at("0xe63d6B308BCe0F6193AeC6b7E6eBa005f41e36AB") # use the address where the contract is deployed
Token.publish_source(token_contract) | [
"tonychew1986@gmail.com"
] | tonychew1986@gmail.com |
fe36ebadacdc7dccccf286e4f14226a17838d061 | e5b4d9e3240b3302db47d770260dd1bcff93d5ef | /BatchShell/batchshell/batchrun.py | e7b2149a60d35511889eb259d80525464d279240 | [
"MIT"
] | permissive | fczj/python-S14 | a28f255b13c15f17d632cf920bca53ed5035752a | 53a1350dae7c797134463c8c42edae44bb5427cb | refs/heads/master | 2021-01-02T09:23:48.195642 | 2017-10-30T12:35:57 | 2017-10-30T12:35:57 | 99,205,729 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,829 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Time : 17-8-26 下午10:08
# @Author : xiongzhibiao
# @Email : 158349411@qqcom
# @File : batchrun.py
# @Software: PyCharm
import sys
sys.path.append('/home/xiongzhibiao/github/python-S14/Common/common')
import paramiko
from multiprocessing import Process,Pool,Lock
import threading
import yaml
import argparse
import sys
import time
from utility import run_time
GROUP_CONF = './../conf/group_conf'
HOST_CONF = './../conf/host_conf'
class Conf():
def __init__(self):
with open(GROUP_CONF,'r') as f:
self.groups = yaml.load(f)
with open(HOST_CONF,'r') as f:
self.hosts= yaml.load(f)
self.all_group = self.groups.keys()
def get_host(self,host_name):
return self.hosts[host_name]
def get_group_all_host(self,group_name):
return self.groups.get(group_name)
class RunCmd():
def __init__(self,user,host,password,port=22):
self.user = user
self.port = port
self.host = host
self.password = password
def run_cmd(self,cmd,lock=''):
ssh = paramiko.SSHClient()
ssh.set_missing_host_key_policy(paramiko.AutoAddPolicy())
ssh.connect(hostname=self.host, port=self.port,
username=self.user,password=self.password)
stdin, stdout, stderr = ssh.exec_command(cmd)
res,err = stdout.read(),stderr.read()
result = res if res else err
ssh.close()
if not lock == '':
lock.acquire()
print ('='*80)
print (self.host)
print(result.decode())
if not lock == '':
lock.release()
def put(self,local,remote):
pass
def get(self,remote,local):
pass
def parse():
parser = argparse.ArgumentParser()
parser.add_argument("-host", "--hostname", nargs='*',dest="host", help="hostname")
parser.add_argument("-g", "--group", nargs='*',dest="group", help="groupname")
parser.add_argument("-cmd", "--command", dest="cmd", help="command")
parser.add_argument("-a", "--action", dest="action", help="action")
args = parser.parse_args()
return args
@run_time
def thread(all_host,cmd):
#多线程版本
lock = threading.Lock()
thread_lst = []
for host in all_host:
conn_info = conf.get_host(host)
cmd_obj = RunCmd(**conn_info)
t = threading.Thread(target=cmd_obj.run_cmd,args=(cmd,lock,))
t.start()
thread_lst.append(t)
for i in thread_lst:
i.join()
def print_log(*args):
print ('='*80)
for i in args:
print (i)
@run_time
def process(all_host,cmd):
lock = Lock()
process_lst = []
for host in all_host:
conn_info = conf.get_host(host)
cmd_obj = RunCmd(**conn_info)
p = Process(target=cmd_obj.run_cmd,args=(cmd,lock,))
p.start()
process_lst.append(p)
for pro in process_lst:
pro.join()
@run_time
def serial(all_host,cmd):
#串行版本
for host in all_host:
conn_info = conf.get_host(host)
cmd_obj = RunCmd(**conn_info)
cmd_obj.run_cmd(cmd)
if __name__ == "__main__":
conf = Conf()
args = parse()
all_host = []
if args.host is not None:
all_host = args.host
if args.group is not None:
groups = args.group
for i in groups:
hosts = conf.get_group_all_host(i)
all_host.extend(hosts)
if args.host is None and args.group is None:
print ('没有可操作对象')
sys.exit(9)
all_host = list(set(all_host))
cmd = args.cmd
#serial(all_host,cmd) #三台主机-1.1秒
thread(all_host,cmd) #三台主机-0.4秒
# process(all_host,cmd) #多线程和多进程时间差不多
| [
"158349411@qq.com"
] | 158349411@qq.com |
03fbeb1450ccc44bd26fc126ce64cfd378980fa0 | 15f321878face2af9317363c5f6de1e5ddd9b749 | /solutions_python/Problem_117/1246.py | 70e284873ba26f9f9d0eb663271fb9c4b2097cdc | [] | no_license | dr-dos-ok/Code_Jam_Webscraper | c06fd59870842664cd79c41eb460a09553e1c80a | 26a35bf114a3aa30fc4c677ef069d95f41665cc0 | refs/heads/master | 2020-04-06T08:17:40.938460 | 2018-10-14T10:12:47 | 2018-10-14T10:12:47 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,615 | py | __author__ = 'joranvar'
__problem__ = 'B'
class Field(object):
def __init__(self, data, height, width):
self.data = data
self.height = height
self.width = width
def is_cuttable(self):
max_x = [max([self.data[y][x] for x in range(self.width)]) for y in range(self.height)]
max_y = [max([self.data[y][x] for y in range(self.height)]) for x in range(self.width)]
for x in range(self.width):
for y in range(self.height):
if self.data[y][x] < min(max_x[y], max_y[x]):
return False
return True
def read_field(f_in, width, height):
field_data = [[int(square) for square in f_in.readline().split()] for line in range(height)]
field = Field(field_data, height, width)
return field
def solve(case, f_in):
N, M = list(map(int, f_in.readline().split()))
field = read_field(f_in, M, N)
if field.is_cuttable(): return ['Case #{}: YES\n'.format(case + 1)]
return ['Case #{}: NO\n'.format(case + 1)]
def open_last_file():
for problem_type in ['-large', '-small-attempt1', '-sample']:
try:
return problem_type, open(__problem__ + problem_type + '.in', 'r')
except FileNotFoundError:
pass
raise FileNotFoundError("No input file found!")
if __name__ == '__main__':
problem_type, f_in = open_last_file()
print (problem_type)
f_out = open(__problem__ + problem_type + '.out', 'w')
T = int(f_in.readline())
for case in range(T):
f_out.writelines(solve(case, f_in))
| [
"miliar1732@gmail.com"
] | miliar1732@gmail.com |
e7573bed0af5548993771609353cbf19e6c6588c | 9dbd215e72f410f810606109632db30556d07f74 | /BuildingSystems/Resources/Scripts/JModelica/Buildings/Examples/BuildingTherma1Zone1DCylinder.py | b5b655006868817fcaba5def05d7357c4c6a71aa | [
"BSD-3-Clause"
] | permissive | TonyJava/BuildingSystems | a7bc822e2277ef87fd7dbce2f2dd685bd311b013 | 56ccd43d1a8de8e2f6dfaed3a2cef9e3edf27b7f | refs/heads/master | 2020-04-24T21:48:34.956614 | 2019-02-22T13:22:34 | 2019-02-22T13:22:34 | 172,290,652 | 1 | 0 | BSD-3-Clause | 2019-02-24T03:18:32 | 2019-02-24T03:18:31 | null | UTF-8 | Python | false | false | 1,817 | py | # paths and info
import os, sys
homeDir = os.environ['HOMEPATH']
jmodDir = os.environ['JMODELICA_HOME']
workDir = "Desktop" # has to be adapted by the user !!!
moLiDir = os.path.join(homeDir, workDir, "BuildingSystems")
# give the path to directory where package.mo is stored
moLibs = [os.path.join(jmodDir, "ThirdParty\MSL\Modelica"),
os.path.join(moLiDir,"BuildingSystems"),
]
print(sys.version)
print(all(os.path.isfile(os.path.join(moLib, "package.mo")) for moLib in moLibs))
print(os.getcwd())
# compile model to fmu
from pymodelica import compile_fmu
model_name = 'BuildingSystems.Buildings.Examples.BuildingThermal1Zone1DCylinder'
my_fmu = compile_fmu(model_name, moLibs)
# simulate the fmu and store results
from pyfmi import load_fmu
myModel = load_fmu(my_fmu)
opts = myModel.simulate_options()
opts['solver'] = "CVode"
opts['ncp'] = 8760
opts['result_handling']="file"
opts["CVode_options"]['discr'] = 'BDF'
opts['CVode_options']['iter'] = 'Newton'
opts['CVode_options']['maxord'] = 5
opts['CVode_options']['atol'] = 1e-5
opts['CVode_options']['rtol'] = 1e-5
res = myModel.simulate(start_time=0.0, final_time=31536000, options=opts)
# plotting of the results
import pylab as P
fig = P.figure(1)
P.clf()
# building
# temperatures
y1 = res['ambient.TAirRef']
y2 = res['building.zone.TAir']
y3 = res['building.zone.TOperative']
t = res['time']
P.subplot(2,1,1)
P.plot(t, y1, t, y2, t, y3)
P.legend(['ambient.TAirRef','building.zone.TAir','building.zone.TOperative'])
P.ylabel('Temperature (K)')
P.xlabel('Time (s)')
# Heating and cooling load
y1 = res['building.zone.Q_flow_heating']
y2 = res['building.zone.Q_flow_cooling']
P.subplot(2,1,2)
P.plot(t, y1, t, y2)
P.legend(['building.zone.Q_flow_heating','building.zone.Q_flow_cooling'])
P.ylabel('power (W)')
P.xlabel('Time (s)')
P.show()
| [
"nytsch@udk-berlin.de"
] | nytsch@udk-berlin.de |
ca99b5cf3ba81bd26679882f2f553c50dc8dabe1 | eb40dce4039d528b9cd06dbeda75da09d09d7fc5 | /need_install/Django-1.8.17/django/db/models/fields/__init__.py | 8bba4cb3216e4eef1a0fcb5822bd70cfc3c685c8 | [
"Apache-2.0",
"BSD-3-Clause"
] | permissive | MulticsYin/MulticsSH | 39b62189446787c7f0f037b1640c9c780bd1dddd | 5837a0bff0e7da0e8535e4e0b31ef6baf24274b4 | refs/heads/master | 2021-08-28T07:53:51.759679 | 2017-12-11T15:31:03 | 2017-12-11T15:31:03 | 82,428,902 | 4 | 2 | null | null | null | null | UTF-8 | Python | false | false | 89,179 | py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
import collections
import copy
import datetime
import decimal
import math
import uuid
import warnings
from base64 import b64decode, b64encode
from itertools import tee
from django.apps import apps
from django.db import connection
from django.db.models.lookups import default_lookups, RegisterLookupMixin
from django.db.models.query_utils import QueryWrapper
from django.conf import settings
from django import forms
from django.core import exceptions, validators, checks
from django.utils.datastructures import DictWrapper
from django.utils.dateparse import parse_date, parse_datetime, parse_time, parse_duration
from django.utils.duration import duration_string
from django.utils.functional import cached_property, curry, total_ordering, Promise
from django.utils.text import capfirst
from django.utils import timezone
from django.utils.translation import ugettext_lazy as _
from django.utils.encoding import (smart_text, force_text, force_bytes,
python_2_unicode_compatible)
from django.utils.ipv6 import clean_ipv6_address
from django.utils import six
from django.utils.itercompat import is_iterable
# When the _meta object was formalized, this exception was moved to
# django.core.exceptions. It is retained here for backwards compatibility
# purposes.
from django.core.exceptions import FieldDoesNotExist # NOQA
# Avoid "TypeError: Item in ``from list'' not a string" -- unicode_literals
# makes these strings unicode
__all__ = [str(x) for x in (
'AutoField', 'BLANK_CHOICE_DASH', 'BigIntegerField', 'BinaryField',
'BooleanField', 'CharField', 'CommaSeparatedIntegerField', 'DateField',
'DateTimeField', 'DecimalField', 'DurationField', 'EmailField', 'Empty',
'Field', 'FieldDoesNotExist', 'FilePathField', 'FloatField',
'GenericIPAddressField', 'IPAddressField', 'IntegerField', 'NOT_PROVIDED',
'NullBooleanField', 'PositiveIntegerField', 'PositiveSmallIntegerField',
'SlugField', 'SmallIntegerField', 'TextField', 'TimeField', 'URLField',
'UUIDField',
)]
class Empty(object):
pass
class NOT_PROVIDED:
pass
# The values to use for "blank" in SelectFields. Will be appended to the start
# of most "choices" lists.
BLANK_CHOICE_DASH = [("", "---------")]
def _load_field(app_label, model_name, field_name):
return apps.get_model(app_label, model_name)._meta.get_field(field_name)
# A guide to Field parameters:
#
# * name: The name of the field specified in the model.
# * attname: The attribute to use on the model object. This is the same as
# "name", except in the case of ForeignKeys, where "_id" is
# appended.
# * db_column: The db_column specified in the model (or None).
# * column: The database column for this field. This is the same as
# "attname", except if db_column is specified.
#
# Code that introspects values, or does other dynamic things, should use
# attname. For example, this gets the primary key value of object "obj":
#
# getattr(obj, opts.pk.attname)
def _empty(of_cls):
new = Empty()
new.__class__ = of_cls
return new
@total_ordering
@python_2_unicode_compatible
class Field(RegisterLookupMixin):
"""Base class for all field types"""
# Designates whether empty strings fundamentally are allowed at the
# database level.
empty_strings_allowed = True
empty_values = list(validators.EMPTY_VALUES)
# These track each time a Field instance is created. Used to retain order.
# The auto_creation_counter is used for fields that Django implicitly
# creates, creation_counter is used for all user-specified fields.
creation_counter = 0
auto_creation_counter = -1
default_validators = [] # Default set of validators
default_error_messages = {
'invalid_choice': _('Value %(value)r is not a valid choice.'),
'null': _('This field cannot be null.'),
'blank': _('This field cannot be blank.'),
'unique': _('%(model_name)s with this %(field_label)s '
'already exists.'),
# Translators: The 'lookup_type' is one of 'date', 'year' or 'month'.
# Eg: "Title must be unique for pub_date year"
'unique_for_date': _("%(field_label)s must be unique for "
"%(date_field_label)s %(lookup_type)s."),
}
class_lookups = default_lookups.copy()
system_check_deprecated_details = None
system_check_removed_details = None
# Field flags
hidden = False
many_to_many = None
many_to_one = None
one_to_many = None
one_to_one = None
related_model = None
# Generic field type description, usually overridden by subclasses
def _description(self):
return _('Field of type: %(field_type)s') % {
'field_type': self.__class__.__name__
}
description = property(_description)
def __init__(self, verbose_name=None, name=None, primary_key=False,
max_length=None, unique=False, blank=False, null=False,
db_index=False, rel=None, default=NOT_PROVIDED, editable=True,
serialize=True, unique_for_date=None, unique_for_month=None,
unique_for_year=None, choices=None, help_text='', db_column=None,
db_tablespace=None, auto_created=False, validators=[],
error_messages=None):
self.name = name
self.verbose_name = verbose_name # May be set by set_attributes_from_name
self._verbose_name = verbose_name # Store original for deconstruction
self.primary_key = primary_key
self.max_length, self._unique = max_length, unique
self.blank, self.null = blank, null
self.rel = rel
self.is_relation = self.rel is not None
self.default = default
self.editable = editable
self.serialize = serialize
self.unique_for_date = unique_for_date
self.unique_for_month = unique_for_month
self.unique_for_year = unique_for_year
self._choices = choices or []
self.help_text = help_text
self.db_column = db_column
self.db_tablespace = db_tablespace or settings.DEFAULT_INDEX_TABLESPACE
self.auto_created = auto_created
# Set db_index to True if the field has a relationship and doesn't
# explicitly set db_index.
self.db_index = db_index
# Adjust the appropriate creation counter, and save our local copy.
if auto_created:
self.creation_counter = Field.auto_creation_counter
Field.auto_creation_counter -= 1
else:
self.creation_counter = Field.creation_counter
Field.creation_counter += 1
self._validators = validators # Store for deconstruction later
messages = {}
for c in reversed(self.__class__.__mro__):
messages.update(getattr(c, 'default_error_messages', {}))
messages.update(error_messages or {})
self._error_messages = error_messages # Store for deconstruction later
self.error_messages = messages
def __str__(self):
""" Return "app_label.model_label.field_name". """
model = self.model
app = model._meta.app_label
return '%s.%s.%s' % (app, model._meta.object_name, self.name)
def __repr__(self):
"""
Displays the module, class and name of the field.
"""
path = '%s.%s' % (self.__class__.__module__, self.__class__.__name__)
name = getattr(self, 'name', None)
if name is not None:
return '<%s: %s>' % (path, name)
return '<%s>' % path
def check(self, **kwargs):
errors = []
errors.extend(self._check_field_name())
errors.extend(self._check_choices())
errors.extend(self._check_db_index())
errors.extend(self._check_null_allowed_for_primary_keys())
errors.extend(self._check_backend_specific_checks(**kwargs))
errors.extend(self._check_deprecation_details())
return errors
def _check_field_name(self):
""" Check if field name is valid, i.e. 1) does not end with an
underscore, 2) does not contain "__" and 3) is not "pk". """
if self.name.endswith('_'):
return [
checks.Error(
'Field names must not end with an underscore.',
hint=None,
obj=self,
id='fields.E001',
)
]
elif '__' in self.name:
return [
checks.Error(
'Field names must not contain "__".',
hint=None,
obj=self,
id='fields.E002',
)
]
elif self.name == 'pk':
return [
checks.Error(
"'pk' is a reserved word that cannot be used as a field name.",
hint=None,
obj=self,
id='fields.E003',
)
]
else:
return []
def _check_choices(self):
if self.choices:
if (isinstance(self.choices, six.string_types) or
not is_iterable(self.choices)):
return [
checks.Error(
"'choices' must be an iterable (e.g., a list or tuple).",
hint=None,
obj=self,
id='fields.E004',
)
]
elif any(isinstance(choice, six.string_types) or
not is_iterable(choice) or len(choice) != 2
for choice in self.choices):
return [
checks.Error(
("'choices' must be an iterable containing "
"(actual value, human readable name) tuples."),
hint=None,
obj=self,
id='fields.E005',
)
]
else:
return []
else:
return []
def _check_db_index(self):
if self.db_index not in (None, True, False):
return [
checks.Error(
"'db_index' must be None, True or False.",
hint=None,
obj=self,
id='fields.E006',
)
]
else:
return []
def _check_null_allowed_for_primary_keys(self):
if (self.primary_key and self.null and
not connection.features.interprets_empty_strings_as_nulls):
# We cannot reliably check this for backends like Oracle which
# consider NULL and '' to be equal (and thus set up
# character-based fields a little differently).
return [
checks.Error(
'Primary keys must not have null=True.',
hint=('Set null=False on the field, or '
'remove primary_key=True argument.'),
obj=self,
id='fields.E007',
)
]
else:
return []
def _check_backend_specific_checks(self, **kwargs):
return connection.validation.check_field(self, **kwargs)
def _check_deprecation_details(self):
if self.system_check_removed_details is not None:
return [
checks.Error(
self.system_check_removed_details.get(
'msg',
'%s has been removed except for support in historical '
'migrations.' % self.__class__.__name__
),
hint=self.system_check_removed_details.get('hint'),
obj=self,
id=self.system_check_removed_details.get('id', 'fields.EXXX'),
)
]
elif self.system_check_deprecated_details is not None:
return [
checks.Warning(
self.system_check_deprecated_details.get(
'msg',
'%s has been deprecated.' % self.__class__.__name__
),
hint=self.system_check_deprecated_details.get('hint'),
obj=self,
id=self.system_check_deprecated_details.get('id', 'fields.WXXX'),
)
]
return []
def get_col(self, alias, output_field=None):
if output_field is None:
output_field = self
if alias != self.model._meta.db_table or output_field != self:
from django.db.models.expressions import Col
return Col(alias, self, output_field)
else:
return self.cached_col
@cached_property
def cached_col(self):
from django.db.models.expressions import Col
return Col(self.model._meta.db_table, self)
def select_format(self, compiler, sql, params):
"""
Custom format for select clauses. For example, GIS columns need to be
selected as AsText(table.col) on MySQL as the table.col data can't be used
by Django.
"""
return sql, params
def deconstruct(self):
"""
Returns enough information to recreate the field as a 4-tuple:
* The name of the field on the model, if contribute_to_class has been run
* The import path of the field, including the class: django.db.models.IntegerField
This should be the most portable version, so less specific may be better.
* A list of positional arguments
* A dict of keyword arguments
Note that the positional or keyword arguments must contain values of the
following types (including inner values of collection types):
* None, bool, str, unicode, int, long, float, complex, set, frozenset, list, tuple, dict
* UUID
* datetime.datetime (naive), datetime.date
* top-level classes, top-level functions - will be referenced by their full import path
* Storage instances - these have their own deconstruct() method
This is because the values here must be serialized into a text format
(possibly new Python code, possibly JSON) and these are the only types
with encoding handlers defined.
There's no need to return the exact way the field was instantiated this time,
just ensure that the resulting field is the same - prefer keyword arguments
over positional ones, and omit parameters with their default values.
"""
# Short-form way of fetching all the default parameters
keywords = {}
possibles = {
"verbose_name": None,
"primary_key": False,
"max_length": None,
"unique": False,
"blank": False,
"null": False,
"db_index": False,
"default": NOT_PROVIDED,
"editable": True,
"serialize": True,
"unique_for_date": None,
"unique_for_month": None,
"unique_for_year": None,
"choices": [],
"help_text": '',
"db_column": None,
"db_tablespace": settings.DEFAULT_INDEX_TABLESPACE,
"auto_created": False,
"validators": [],
"error_messages": None,
}
attr_overrides = {
"unique": "_unique",
"choices": "_choices",
"error_messages": "_error_messages",
"validators": "_validators",
"verbose_name": "_verbose_name",
}
equals_comparison = {"choices", "validators", "db_tablespace"}
for name, default in possibles.items():
value = getattr(self, attr_overrides.get(name, name))
# Unroll anything iterable for choices into a concrete list
if name == "choices" and isinstance(value, collections.Iterable):
value = list(value)
# Do correct kind of comparison
if name in equals_comparison:
if value != default:
keywords[name] = value
else:
if value is not default:
keywords[name] = value
# Work out path - we shorten it for known Django core fields
path = "%s.%s" % (self.__class__.__module__, self.__class__.__name__)
if path.startswith("django.db.models.fields.related"):
path = path.replace("django.db.models.fields.related", "django.db.models")
if path.startswith("django.db.models.fields.files"):
path = path.replace("django.db.models.fields.files", "django.db.models")
if path.startswith("django.db.models.fields.proxy"):
path = path.replace("django.db.models.fields.proxy", "django.db.models")
if path.startswith("django.db.models.fields"):
path = path.replace("django.db.models.fields", "django.db.models")
# Return basic info - other fields should override this.
return (
force_text(self.name, strings_only=True),
path,
[],
keywords,
)
def clone(self):
"""
Uses deconstruct() to clone a new copy of this Field.
Will not preserve any class attachments/attribute names.
"""
name, path, args, kwargs = self.deconstruct()
return self.__class__(*args, **kwargs)
def __eq__(self, other):
# Needed for @total_ordering
if isinstance(other, Field):
return self.creation_counter == other.creation_counter
return NotImplemented
def __lt__(self, other):
# This is needed because bisect does not take a comparison function.
if isinstance(other, Field):
return self.creation_counter < other.creation_counter
return NotImplemented
def __hash__(self):
return hash(self.creation_counter)
def __deepcopy__(self, memodict):
# We don't have to deepcopy very much here, since most things are not
# intended to be altered after initial creation.
obj = copy.copy(self)
if self.rel:
obj.rel = copy.copy(self.rel)
if hasattr(self.rel, 'field') and self.rel.field is self:
obj.rel.field = obj
memodict[id(self)] = obj
return obj
def __copy__(self):
# We need to avoid hitting __reduce__, so define this
# slightly weird copy construct.
obj = Empty()
obj.__class__ = self.__class__
obj.__dict__ = self.__dict__.copy()
return obj
def __reduce__(self):
"""
Pickling should return the model._meta.fields instance of the field,
not a new copy of that field. So, we use the app registry to load the
model and then the field back.
"""
if not hasattr(self, 'model'):
# Fields are sometimes used without attaching them to models (for
# example in aggregation). In this case give back a plain field
# instance. The code below will create a new empty instance of
# class self.__class__, then update its dict with self.__dict__
# values - so, this is very close to normal pickle.
return _empty, (self.__class__,), self.__dict__
if self.model._deferred:
# Deferred model will not be found from the app registry. This
# could be fixed by reconstructing the deferred model on unpickle.
raise RuntimeError("Fields of deferred models can't be reduced")
return _load_field, (self.model._meta.app_label, self.model._meta.object_name,
self.name)
def get_pk_value_on_save(self, instance):
"""
Hook to generate new PK values on save. This method is called when
saving instances with no primary key value set. If this method returns
something else than None, then the returned value is used when saving
the new instance.
"""
if self.default:
return self.get_default()
return None
def to_python(self, value):
"""
Converts the input value into the expected Python data type, raising
django.core.exceptions.ValidationError if the data can't be converted.
Returns the converted value. Subclasses should override this.
"""
return value
@cached_property
def validators(self):
# Some validators can't be created at field initialization time.
# This method provides a way to delay their creation until required.
return self.default_validators + self._validators
def run_validators(self, value):
if value in self.empty_values:
return
errors = []
for v in self.validators:
try:
v(value)
except exceptions.ValidationError as e:
if hasattr(e, 'code') and e.code in self.error_messages:
e.message = self.error_messages[e.code]
errors.extend(e.error_list)
if errors:
raise exceptions.ValidationError(errors)
def validate(self, value, model_instance):
"""
Validates value and throws ValidationError. Subclasses should override
this to provide validation logic.
"""
if not self.editable:
# Skip validation for non-editable fields.
return
if self._choices and value not in self.empty_values:
for option_key, option_value in self.choices:
if isinstance(option_value, (list, tuple)):
# This is an optgroup, so look inside the group for
# options.
for optgroup_key, optgroup_value in option_value:
if value == optgroup_key:
return
elif value == option_key:
return
raise exceptions.ValidationError(
self.error_messages['invalid_choice'],
code='invalid_choice',
params={'value': value},
)
if value is None and not self.null:
raise exceptions.ValidationError(self.error_messages['null'], code='null')
if not self.blank and value in self.empty_values:
raise exceptions.ValidationError(self.error_messages['blank'], code='blank')
def clean(self, value, model_instance):
"""
Convert the value's type and run validation. Validation errors
from to_python and validate are propagated. The correct value is
returned if no error is raised.
"""
value = self.to_python(value)
self.validate(value, model_instance)
self.run_validators(value)
return value
def db_type(self, connection):
"""
Returns the database column data type for this field, for the provided
connection.
"""
# The default implementation of this method looks at the
# backend-specific data_types dictionary, looking up the field by its
# "internal type".
#
# A Field class can implement the get_internal_type() method to specify
# which *preexisting* Django Field class it's most similar to -- i.e.,
# a custom field might be represented by a TEXT column type, which is
# the same as the TextField Django field type, which means the custom
# field's get_internal_type() returns 'TextField'.
#
# But the limitation of the get_internal_type() / data_types approach
# is that it cannot handle database column types that aren't already
# mapped to one of the built-in Django field types. In this case, you
# can implement db_type() instead of get_internal_type() to specify
# exactly which wacky database column type you want to use.
data = DictWrapper(self.__dict__, connection.ops.quote_name, "qn_")
try:
return connection.data_types[self.get_internal_type()] % data
except KeyError:
return None
def db_parameters(self, connection):
"""
Extension of db_type(), providing a range of different return
values (type, checks).
This will look at db_type(), allowing custom model fields to override it.
"""
data = DictWrapper(self.__dict__, connection.ops.quote_name, "qn_")
type_string = self.db_type(connection)
try:
check_string = connection.data_type_check_constraints[self.get_internal_type()] % data
except KeyError:
check_string = None
return {
"type": type_string,
"check": check_string,
}
def db_type_suffix(self, connection):
return connection.data_types_suffix.get(self.get_internal_type())
def get_db_converters(self, connection):
if hasattr(self, 'from_db_value'):
return [self.from_db_value]
return []
@property
def unique(self):
return self._unique or self.primary_key
def set_attributes_from_name(self, name):
if not self.name:
self.name = name
self.attname, self.column = self.get_attname_column()
self.concrete = self.column is not None
if self.verbose_name is None and self.name:
self.verbose_name = self.name.replace('_', ' ')
def contribute_to_class(self, cls, name, virtual_only=False):
self.set_attributes_from_name(name)
self.model = cls
if virtual_only:
cls._meta.add_field(self, virtual=True)
else:
cls._meta.add_field(self)
if self.choices:
setattr(cls, 'get_%s_display' % self.name,
curry(cls._get_FIELD_display, field=self))
def get_attname(self):
return self.name
def get_attname_column(self):
attname = self.get_attname()
column = self.db_column or attname
return attname, column
def get_cache_name(self):
return '_%s_cache' % self.name
def get_internal_type(self):
return self.__class__.__name__
def pre_save(self, model_instance, add):
"""
Returns field's value just before saving.
"""
return getattr(model_instance, self.attname)
def get_prep_value(self, value):
"""
Perform preliminary non-db specific value checks and conversions.
"""
if isinstance(value, Promise):
value = value._proxy____cast()
return value
def get_db_prep_value(self, value, connection, prepared=False):
"""Returns field's value prepared for interacting with the database
backend.
Used by the default implementations of ``get_db_prep_save``and
`get_db_prep_lookup```
"""
if not prepared:
value = self.get_prep_value(value)
return value
def get_db_prep_save(self, value, connection):
"""
Returns field's value prepared for saving into a database.
"""
return self.get_db_prep_value(value, connection=connection,
prepared=False)
def get_prep_lookup(self, lookup_type, value):
"""
Perform preliminary non-db specific lookup checks and conversions
"""
if hasattr(value, '_prepare'):
return value._prepare()
if lookup_type in {
'iexact', 'contains', 'icontains',
'startswith', 'istartswith', 'endswith', 'iendswith',
'month', 'day', 'week_day', 'hour', 'minute', 'second',
'isnull', 'search', 'regex', 'iregex',
}:
return value
elif lookup_type in ('exact', 'gt', 'gte', 'lt', 'lte'):
return self.get_prep_value(value)
elif lookup_type in ('range', 'in'):
return [self.get_prep_value(v) for v in value]
elif lookup_type == 'year':
try:
return int(value)
except ValueError:
raise ValueError("The __year lookup type requires an integer "
"argument")
return self.get_prep_value(value)
def get_db_prep_lookup(self, lookup_type, value, connection,
prepared=False):
"""
Returns field's value prepared for database lookup.
"""
if not prepared:
value = self.get_prep_lookup(lookup_type, value)
prepared = True
if hasattr(value, 'get_compiler'):
value = value.get_compiler(connection=connection)
if hasattr(value, 'as_sql') or hasattr(value, '_as_sql'):
# If the value has a relabeled_clone method it means the
# value will be handled later on.
if hasattr(value, 'relabeled_clone'):
return value
if hasattr(value, 'as_sql'):
sql, params = value.as_sql()
else:
sql, params = value._as_sql(connection=connection)
return QueryWrapper(('(%s)' % sql), params)
if lookup_type in ('month', 'day', 'week_day', 'hour', 'minute',
'second', 'search', 'regex', 'iregex', 'contains',
'icontains', 'iexact', 'startswith', 'endswith',
'istartswith', 'iendswith'):
return [value]
elif lookup_type in ('exact', 'gt', 'gte', 'lt', 'lte'):
return [self.get_db_prep_value(value, connection=connection,
prepared=prepared)]
elif lookup_type in ('range', 'in'):
return [self.get_db_prep_value(v, connection=connection,
prepared=prepared) for v in value]
elif lookup_type == 'isnull':
return []
elif lookup_type == 'year':
if isinstance(self, DateTimeField):
return connection.ops.year_lookup_bounds_for_datetime_field(value)
elif isinstance(self, DateField):
return connection.ops.year_lookup_bounds_for_date_field(value)
else:
return [value] # this isn't supposed to happen
else:
return [value]
def has_default(self):
"""
Returns a boolean of whether this field has a default value.
"""
return self.default is not NOT_PROVIDED
def get_default(self):
"""
Returns the default value for this field.
"""
if self.has_default():
if callable(self.default):
return self.default()
return self.default
if (not self.empty_strings_allowed or (self.null and
not connection.features.interprets_empty_strings_as_nulls)):
return None
return ""
def get_choices(self, include_blank=True, blank_choice=BLANK_CHOICE_DASH, limit_choices_to=None):
"""Returns choices with a default blank choices included, for use
as SelectField choices for this field."""
blank_defined = False
choices = list(self.choices) if self.choices else []
named_groups = choices and isinstance(choices[0][1], (list, tuple))
if not named_groups:
for choice, __ in choices:
if choice in ('', None):
blank_defined = True
break
first_choice = (blank_choice if include_blank and
not blank_defined else [])
if self.choices:
return first_choice + choices
rel_model = self.rel.to
limit_choices_to = limit_choices_to or self.get_limit_choices_to()
if hasattr(self.rel, 'get_related_field'):
lst = [(getattr(x, self.rel.get_related_field().attname),
smart_text(x))
for x in rel_model._default_manager.complex_filter(
limit_choices_to)]
else:
lst = [(x._get_pk_val(), smart_text(x))
for x in rel_model._default_manager.complex_filter(
limit_choices_to)]
return first_choice + lst
def get_choices_default(self):
return self.get_choices()
def get_flatchoices(self, include_blank=True,
blank_choice=BLANK_CHOICE_DASH):
"""
Returns flattened choices with a default blank choice included.
"""
first_choice = blank_choice if include_blank else []
return first_choice + list(self.flatchoices)
def _get_val_from_obj(self, obj):
if obj is not None:
return getattr(obj, self.attname)
else:
return self.get_default()
def value_to_string(self, obj):
"""
Returns a string value of this field from the passed obj.
This is used by the serialization framework.
"""
return smart_text(self._get_val_from_obj(obj))
def _get_choices(self):
if isinstance(self._choices, collections.Iterator):
choices, self._choices = tee(self._choices)
return choices
else:
return self._choices
choices = property(_get_choices)
def _get_flatchoices(self):
"""Flattened version of choices tuple."""
flat = []
for choice, value in self.choices:
if isinstance(value, (list, tuple)):
flat.extend(value)
else:
flat.append((choice, value))
return flat
flatchoices = property(_get_flatchoices)
def save_form_data(self, instance, data):
setattr(instance, self.name, data)
def formfield(self, form_class=None, choices_form_class=None, **kwargs):
"""
Returns a django.forms.Field instance for this database Field.
"""
defaults = {'required': not self.blank,
'label': capfirst(self.verbose_name),
'help_text': self.help_text}
if self.has_default():
if callable(self.default):
defaults['initial'] = self.default
defaults['show_hidden_initial'] = True
else:
defaults['initial'] = self.get_default()
if self.choices:
# Fields with choices get special treatment.
include_blank = (self.blank or
not (self.has_default() or 'initial' in kwargs))
defaults['choices'] = self.get_choices(include_blank=include_blank)
defaults['coerce'] = self.to_python
if self.null:
defaults['empty_value'] = None
if choices_form_class is not None:
form_class = choices_form_class
else:
form_class = forms.TypedChoiceField
# Many of the subclass-specific formfield arguments (min_value,
# max_value) don't apply for choice fields, so be sure to only pass
# the values that TypedChoiceField will understand.
for k in list(kwargs):
if k not in ('coerce', 'empty_value', 'choices', 'required',
'widget', 'label', 'initial', 'help_text',
'error_messages', 'show_hidden_initial'):
del kwargs[k]
defaults.update(kwargs)
if form_class is None:
form_class = forms.CharField
return form_class(**defaults)
def value_from_object(self, obj):
"""
Returns the value of this field in the given model instance.
"""
return getattr(obj, self.attname)
class AutoField(Field):
description = _("Integer")
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be an integer."),
}
def __init__(self, *args, **kwargs):
kwargs['blank'] = True
super(AutoField, self).__init__(*args, **kwargs)
def check(self, **kwargs):
errors = super(AutoField, self).check(**kwargs)
errors.extend(self._check_primary_key())
return errors
def _check_primary_key(self):
if not self.primary_key:
return [
checks.Error(
'AutoFields must set primary_key=True.',
hint=None,
obj=self,
id='fields.E100',
),
]
else:
return []
def deconstruct(self):
name, path, args, kwargs = super(AutoField, self).deconstruct()
del kwargs['blank']
kwargs['primary_key'] = True
return name, path, args, kwargs
def get_internal_type(self):
return "AutoField"
def to_python(self, value):
if value is None:
return value
try:
return int(value)
except (TypeError, ValueError):
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def validate(self, value, model_instance):
pass
def get_db_prep_value(self, value, connection, prepared=False):
if not prepared:
value = self.get_prep_value(value)
value = connection.ops.validate_autopk_value(value)
return value
def get_prep_value(self, value):
value = super(AutoField, self).get_prep_value(value)
if value is None:
return None
return int(value)
def contribute_to_class(self, cls, name, **kwargs):
assert not cls._meta.has_auto_field, \
"A model can't have more than one AutoField."
super(AutoField, self).contribute_to_class(cls, name, **kwargs)
cls._meta.has_auto_field = True
cls._meta.auto_field = self
def formfield(self, **kwargs):
return None
class BooleanField(Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be either True or False."),
}
description = _("Boolean (Either True or False)")
def __init__(self, *args, **kwargs):
kwargs['blank'] = True
super(BooleanField, self).__init__(*args, **kwargs)
def check(self, **kwargs):
errors = super(BooleanField, self).check(**kwargs)
errors.extend(self._check_null(**kwargs))
return errors
def _check_null(self, **kwargs):
if getattr(self, 'null', False):
return [
checks.Error(
'BooleanFields do not accept null values.',
hint='Use a NullBooleanField instead.',
obj=self,
id='fields.E110',
)
]
else:
return []
def deconstruct(self):
name, path, args, kwargs = super(BooleanField, self).deconstruct()
del kwargs['blank']
return name, path, args, kwargs
def get_internal_type(self):
return "BooleanField"
def to_python(self, value):
if value in (True, False):
# if value is 1 or 0 than it's equal to True or False, but we want
# to return a true bool for semantic reasons.
return bool(value)
if value in ('t', 'True', '1'):
return True
if value in ('f', 'False', '0'):
return False
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def get_prep_lookup(self, lookup_type, value):
# Special-case handling for filters coming from a Web request (e.g. the
# admin interface). Only works for scalar values (not lists). If you're
# passing in a list, you might as well make things the right type when
# constructing the list.
if value in ('1', '0'):
value = bool(int(value))
return super(BooleanField, self).get_prep_lookup(lookup_type, value)
def get_prep_value(self, value):
value = super(BooleanField, self).get_prep_value(value)
if value is None:
return None
return bool(value)
def formfield(self, **kwargs):
# Unlike most fields, BooleanField figures out include_blank from
# self.null instead of self.blank.
if self.choices:
include_blank = not (self.has_default() or 'initial' in kwargs)
defaults = {'choices': self.get_choices(include_blank=include_blank)}
else:
defaults = {'form_class': forms.BooleanField}
defaults.update(kwargs)
return super(BooleanField, self).formfield(**defaults)
class CharField(Field):
description = _("String (up to %(max_length)s)")
def __init__(self, *args, **kwargs):
super(CharField, self).__init__(*args, **kwargs)
self.validators.append(validators.MaxLengthValidator(self.max_length))
def check(self, **kwargs):
errors = super(CharField, self).check(**kwargs)
errors.extend(self._check_max_length_attribute(**kwargs))
return errors
def _check_max_length_attribute(self, **kwargs):
try:
max_length = int(self.max_length)
if max_length <= 0:
raise ValueError()
except TypeError:
return [
checks.Error(
"CharFields must define a 'max_length' attribute.",
hint=None,
obj=self,
id='fields.E120',
)
]
except ValueError:
return [
checks.Error(
"'max_length' must be a positive integer.",
hint=None,
obj=self,
id='fields.E121',
)
]
else:
return []
def get_internal_type(self):
return "CharField"
def to_python(self, value):
if isinstance(value, six.string_types) or value is None:
return value
return smart_text(value)
def get_prep_value(self, value):
value = super(CharField, self).get_prep_value(value)
return self.to_python(value)
def formfield(self, **kwargs):
# Passing max_length to forms.CharField means that the value's length
# will be validated twice. This is considered acceptable since we want
# the value in the form field (to pass into widget for example).
defaults = {'max_length': self.max_length}
defaults.update(kwargs)
return super(CharField, self).formfield(**defaults)
# TODO: Maybe move this into contrib, because it's specialized.
class CommaSeparatedIntegerField(CharField):
default_validators = [validators.validate_comma_separated_integer_list]
description = _("Comma-separated integers")
def formfield(self, **kwargs):
defaults = {
'error_messages': {
'invalid': _('Enter only digits separated by commas.'),
}
}
defaults.update(kwargs)
return super(CommaSeparatedIntegerField, self).formfield(**defaults)
class DateTimeCheckMixin(object):
def check(self, **kwargs):
errors = super(DateTimeCheckMixin, self).check(**kwargs)
errors.extend(self._check_mutually_exclusive_options())
errors.extend(self._check_fix_default_value())
return errors
def _check_mutually_exclusive_options(self):
# auto_now, auto_now_add, and default are mutually exclusive
# options. The use of more than one of these options together
# will trigger an Error
mutually_exclusive_options = [self.auto_now_add, self.auto_now,
self.has_default()]
enabled_options = [option not in (None, False)
for option in mutually_exclusive_options].count(True)
if enabled_options > 1:
return [
checks.Error(
"The options auto_now, auto_now_add, and default "
"are mutually exclusive. Only one of these options "
"may be present.",
hint=None,
obj=self,
id='fields.E160',
)
]
else:
return []
def _check_fix_default_value(self):
return []
class DateField(DateTimeCheckMixin, Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value has an invalid date format. It must be "
"in YYYY-MM-DD format."),
'invalid_date': _("'%(value)s' value has the correct format (YYYY-MM-DD) "
"but it is an invalid date."),
}
description = _("Date (without time)")
def __init__(self, verbose_name=None, name=None, auto_now=False,
auto_now_add=False, **kwargs):
self.auto_now, self.auto_now_add = auto_now, auto_now_add
if auto_now or auto_now_add:
kwargs['editable'] = False
kwargs['blank'] = True
super(DateField, self).__init__(verbose_name, name, **kwargs)
def _check_fix_default_value(self):
"""
Adds a warning to the checks framework stating, that using an actual
date or datetime value is probably wrong; it's only being evaluated on
server start-up.
For details see ticket #21905
"""
if not self.has_default():
return []
now = timezone.now()
if not timezone.is_naive(now):
now = timezone.make_naive(now, timezone.utc)
value = self.default
if isinstance(value, datetime.datetime):
if not timezone.is_naive(value):
value = timezone.make_naive(value, timezone.utc)
value = value.date()
elif isinstance(value, datetime.date):
# Nothing to do, as dates don't have tz information
pass
else:
# No explicit date / datetime value -- no checks necessary
return []
offset = datetime.timedelta(days=1)
lower = (now - offset).date()
upper = (now + offset).date()
if lower <= value <= upper:
return [
checks.Warning(
'Fixed default value provided.',
hint='It seems you set a fixed date / time / datetime '
'value as default for this field. This may not be '
'what you want. If you want to have the current date '
'as default, use `django.utils.timezone.now`',
obj=self,
id='fields.W161',
)
]
return []
def deconstruct(self):
name, path, args, kwargs = super(DateField, self).deconstruct()
if self.auto_now:
kwargs['auto_now'] = True
if self.auto_now_add:
kwargs['auto_now_add'] = True
if self.auto_now or self.auto_now_add:
del kwargs['editable']
del kwargs['blank']
return name, path, args, kwargs
def get_internal_type(self):
return "DateField"
def to_python(self, value):
if value is None:
return value
if isinstance(value, datetime.datetime):
if settings.USE_TZ and timezone.is_aware(value):
# Convert aware datetimes to the default time zone
# before casting them to dates (#17742).
default_timezone = timezone.get_default_timezone()
value = timezone.make_naive(value, default_timezone)
return value.date()
if isinstance(value, datetime.date):
return value
try:
parsed = parse_date(value)
if parsed is not None:
return parsed
except ValueError:
raise exceptions.ValidationError(
self.error_messages['invalid_date'],
code='invalid_date',
params={'value': value},
)
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def pre_save(self, model_instance, add):
if self.auto_now or (self.auto_now_add and add):
value = datetime.date.today()
setattr(model_instance, self.attname, value)
return value
else:
return super(DateField, self).pre_save(model_instance, add)
def contribute_to_class(self, cls, name, **kwargs):
super(DateField, self).contribute_to_class(cls, name, **kwargs)
if not self.null:
setattr(cls, 'get_next_by_%s' % self.name,
curry(cls._get_next_or_previous_by_FIELD, field=self,
is_next=True))
setattr(cls, 'get_previous_by_%s' % self.name,
curry(cls._get_next_or_previous_by_FIELD, field=self,
is_next=False))
def get_prep_lookup(self, lookup_type, value):
# For dates lookups, convert the value to an int
# so the database backend always sees a consistent type.
if lookup_type in ('month', 'day', 'week_day', 'hour', 'minute', 'second'):
return int(value)
return super(DateField, self).get_prep_lookup(lookup_type, value)
def get_prep_value(self, value):
value = super(DateField, self).get_prep_value(value)
return self.to_python(value)
def get_db_prep_value(self, value, connection, prepared=False):
# Casts dates into the format expected by the backend
if not prepared:
value = self.get_prep_value(value)
return connection.ops.value_to_db_date(value)
def value_to_string(self, obj):
val = self._get_val_from_obj(obj)
return '' if val is None else val.isoformat()
def formfield(self, **kwargs):
defaults = {'form_class': forms.DateField}
defaults.update(kwargs)
return super(DateField, self).formfield(**defaults)
class DateTimeField(DateField):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value has an invalid format. It must be in "
"YYYY-MM-DD HH:MM[:ss[.uuuuuu]][TZ] format."),
'invalid_date': _("'%(value)s' value has the correct format "
"(YYYY-MM-DD) but it is an invalid date."),
'invalid_datetime': _("'%(value)s' value has the correct format "
"(YYYY-MM-DD HH:MM[:ss[.uuuuuu]][TZ]) "
"but it is an invalid date/time."),
}
description = _("Date (with time)")
# __init__ is inherited from DateField
def _check_fix_default_value(self):
"""
Adds a warning to the checks framework stating, that using an actual
date or datetime value is probably wrong; it's only being evaluated on
server start-up.
For details see ticket #21905
"""
if not self.has_default():
return []
now = timezone.now()
if not timezone.is_naive(now):
now = timezone.make_naive(now, timezone.utc)
value = self.default
if isinstance(value, datetime.datetime):
second_offset = datetime.timedelta(seconds=10)
lower = now - second_offset
upper = now + second_offset
if timezone.is_aware(value):
value = timezone.make_naive(value, timezone.utc)
elif isinstance(value, datetime.date):
second_offset = datetime.timedelta(seconds=10)
lower = now - second_offset
lower = datetime.datetime(lower.year, lower.month, lower.day)
upper = now + second_offset
upper = datetime.datetime(upper.year, upper.month, upper.day)
value = datetime.datetime(value.year, value.month, value.day)
else:
# No explicit date / datetime value -- no checks necessary
return []
if lower <= value <= upper:
return [
checks.Warning(
'Fixed default value provided.',
hint='It seems you set a fixed date / time / datetime '
'value as default for this field. This may not be '
'what you want. If you want to have the current date '
'as default, use `django.utils.timezone.now`',
obj=self,
id='fields.W161',
)
]
return []
def get_internal_type(self):
return "DateTimeField"
def to_python(self, value):
if value is None:
return value
if isinstance(value, datetime.datetime):
return value
if isinstance(value, datetime.date):
value = datetime.datetime(value.year, value.month, value.day)
if settings.USE_TZ:
# For backwards compatibility, interpret naive datetimes in
# local time. This won't work during DST change, but we can't
# do much about it, so we let the exceptions percolate up the
# call stack.
warnings.warn("DateTimeField %s.%s received a naive datetime "
"(%s) while time zone support is active." %
(self.model.__name__, self.name, value),
RuntimeWarning)
default_timezone = timezone.get_default_timezone()
value = timezone.make_aware(value, default_timezone)
return value
try:
parsed = parse_datetime(value)
if parsed is not None:
return parsed
except ValueError:
raise exceptions.ValidationError(
self.error_messages['invalid_datetime'],
code='invalid_datetime',
params={'value': value},
)
try:
parsed = parse_date(value)
if parsed is not None:
return datetime.datetime(parsed.year, parsed.month, parsed.day)
except ValueError:
raise exceptions.ValidationError(
self.error_messages['invalid_date'],
code='invalid_date',
params={'value': value},
)
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def pre_save(self, model_instance, add):
if self.auto_now or (self.auto_now_add and add):
value = timezone.now()
setattr(model_instance, self.attname, value)
return value
else:
return super(DateTimeField, self).pre_save(model_instance, add)
# contribute_to_class is inherited from DateField, it registers
# get_next_by_FOO and get_prev_by_FOO
# get_prep_lookup is inherited from DateField
def get_prep_value(self, value):
value = super(DateTimeField, self).get_prep_value(value)
value = self.to_python(value)
if value is not None and settings.USE_TZ and timezone.is_naive(value):
# For backwards compatibility, interpret naive datetimes in local
# time. This won't work during DST change, but we can't do much
# about it, so we let the exceptions percolate up the call stack.
try:
name = '%s.%s' % (self.model.__name__, self.name)
except AttributeError:
name = '(unbound)'
warnings.warn("DateTimeField %s received a naive datetime (%s)"
" while time zone support is active." %
(name, value),
RuntimeWarning)
default_timezone = timezone.get_default_timezone()
value = timezone.make_aware(value, default_timezone)
return value
def get_db_prep_value(self, value, connection, prepared=False):
# Casts datetimes into the format expected by the backend
if not prepared:
value = self.get_prep_value(value)
return connection.ops.value_to_db_datetime(value)
def value_to_string(self, obj):
val = self._get_val_from_obj(obj)
return '' if val is None else val.isoformat()
def formfield(self, **kwargs):
defaults = {'form_class': forms.DateTimeField}
defaults.update(kwargs)
return super(DateTimeField, self).formfield(**defaults)
class DecimalField(Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be a decimal number."),
}
description = _("Decimal number")
def __init__(self, verbose_name=None, name=None, max_digits=None,
decimal_places=None, **kwargs):
self.max_digits, self.decimal_places = max_digits, decimal_places
super(DecimalField, self).__init__(verbose_name, name, **kwargs)
def check(self, **kwargs):
errors = super(DecimalField, self).check(**kwargs)
digits_errors = self._check_decimal_places()
digits_errors.extend(self._check_max_digits())
if not digits_errors:
errors.extend(self._check_decimal_places_and_max_digits(**kwargs))
else:
errors.extend(digits_errors)
return errors
def _check_decimal_places(self):
try:
decimal_places = int(self.decimal_places)
if decimal_places < 0:
raise ValueError()
except TypeError:
return [
checks.Error(
"DecimalFields must define a 'decimal_places' attribute.",
hint=None,
obj=self,
id='fields.E130',
)
]
except ValueError:
return [
checks.Error(
"'decimal_places' must be a non-negative integer.",
hint=None,
obj=self,
id='fields.E131',
)
]
else:
return []
def _check_max_digits(self):
try:
max_digits = int(self.max_digits)
if max_digits <= 0:
raise ValueError()
except TypeError:
return [
checks.Error(
"DecimalFields must define a 'max_digits' attribute.",
hint=None,
obj=self,
id='fields.E132',
)
]
except ValueError:
return [
checks.Error(
"'max_digits' must be a positive integer.",
hint=None,
obj=self,
id='fields.E133',
)
]
else:
return []
def _check_decimal_places_and_max_digits(self, **kwargs):
if int(self.decimal_places) > int(self.max_digits):
return [
checks.Error(
"'max_digits' must be greater or equal to 'decimal_places'.",
hint=None,
obj=self,
id='fields.E134',
)
]
return []
def deconstruct(self):
name, path, args, kwargs = super(DecimalField, self).deconstruct()
if self.max_digits is not None:
kwargs['max_digits'] = self.max_digits
if self.decimal_places is not None:
kwargs['decimal_places'] = self.decimal_places
return name, path, args, kwargs
def get_internal_type(self):
return "DecimalField"
def to_python(self, value):
if value is None:
return value
try:
return decimal.Decimal(value)
except decimal.InvalidOperation:
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def _format(self, value):
if isinstance(value, six.string_types):
return value
else:
return self.format_number(value)
def format_number(self, value):
"""
Formats a number into a string with the requisite number of digits and
decimal places.
"""
# Method moved to django.db.backends.utils.
#
# It is preserved because it is used by the oracle backend
# (django.db.backends.oracle.query), and also for
# backwards-compatibility with any external code which may have used
# this method.
from django.db.backends import utils
return utils.format_number(value, self.max_digits, self.decimal_places)
def get_db_prep_save(self, value, connection):
return connection.ops.value_to_db_decimal(self.to_python(value),
self.max_digits, self.decimal_places)
def get_prep_value(self, value):
value = super(DecimalField, self).get_prep_value(value)
return self.to_python(value)
def formfield(self, **kwargs):
defaults = {
'max_digits': self.max_digits,
'decimal_places': self.decimal_places,
'form_class': forms.DecimalField,
}
defaults.update(kwargs)
return super(DecimalField, self).formfield(**defaults)
class DurationField(Field):
"""Stores timedelta objects.
Uses interval on postgres, INVERAL DAY TO SECOND on Oracle, and bigint of
microseconds on other databases.
"""
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value has an invalid format. It must be in "
"[DD] [HH:[MM:]]ss[.uuuuuu] format.")
}
description = _("Duration")
def get_internal_type(self):
return "DurationField"
def to_python(self, value):
if value is None:
return value
if isinstance(value, datetime.timedelta):
return value
try:
parsed = parse_duration(value)
except ValueError:
pass
else:
if parsed is not None:
return parsed
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def get_db_prep_value(self, value, connection, prepared=False):
if connection.features.has_native_duration_field:
return value
if value is None:
return None
# Discard any fractional microseconds due to floating point arithmetic.
return int(round(value.total_seconds() * 1000000))
def get_db_converters(self, connection):
converters = []
if not connection.features.has_native_duration_field:
converters.append(connection.ops.convert_durationfield_value)
return converters + super(DurationField, self).get_db_converters(connection)
def value_to_string(self, obj):
val = self._get_val_from_obj(obj)
return '' if val is None else duration_string(val)
def formfield(self, **kwargs):
defaults = {
'form_class': forms.DurationField,
}
defaults.update(kwargs)
return super(DurationField, self).formfield(**defaults)
class EmailField(CharField):
default_validators = [validators.validate_email]
description = _("Email address")
def __init__(self, *args, **kwargs):
# max_length=254 to be compliant with RFCs 3696 and 5321
kwargs['max_length'] = kwargs.get('max_length', 254)
super(EmailField, self).__init__(*args, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(EmailField, self).deconstruct()
# We do not exclude max_length if it matches default as we want to change
# the default in future.
return name, path, args, kwargs
def formfield(self, **kwargs):
# As with CharField, this will cause email validation to be performed
# twice.
defaults = {
'form_class': forms.EmailField,
}
defaults.update(kwargs)
return super(EmailField, self).formfield(**defaults)
class FilePathField(Field):
description = _("File path")
def __init__(self, verbose_name=None, name=None, path='', match=None,
recursive=False, allow_files=True, allow_folders=False, **kwargs):
self.path, self.match, self.recursive = path, match, recursive
self.allow_files, self.allow_folders = allow_files, allow_folders
kwargs['max_length'] = kwargs.get('max_length', 100)
super(FilePathField, self).__init__(verbose_name, name, **kwargs)
def check(self, **kwargs):
errors = super(FilePathField, self).check(**kwargs)
errors.extend(self._check_allowing_files_or_folders(**kwargs))
return errors
def _check_allowing_files_or_folders(self, **kwargs):
if not self.allow_files and not self.allow_folders:
return [
checks.Error(
"FilePathFields must have either 'allow_files' or 'allow_folders' set to True.",
hint=None,
obj=self,
id='fields.E140',
)
]
return []
def deconstruct(self):
name, path, args, kwargs = super(FilePathField, self).deconstruct()
if self.path != '':
kwargs['path'] = self.path
if self.match is not None:
kwargs['match'] = self.match
if self.recursive is not False:
kwargs['recursive'] = self.recursive
if self.allow_files is not True:
kwargs['allow_files'] = self.allow_files
if self.allow_folders is not False:
kwargs['allow_folders'] = self.allow_folders
if kwargs.get("max_length", None) == 100:
del kwargs["max_length"]
return name, path, args, kwargs
def get_prep_value(self, value):
value = super(FilePathField, self).get_prep_value(value)
if value is None:
return None
return six.text_type(value)
def formfield(self, **kwargs):
defaults = {
'path': self.path,
'match': self.match,
'recursive': self.recursive,
'form_class': forms.FilePathField,
'allow_files': self.allow_files,
'allow_folders': self.allow_folders,
}
defaults.update(kwargs)
return super(FilePathField, self).formfield(**defaults)
def get_internal_type(self):
return "FilePathField"
class FloatField(Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be a float."),
}
description = _("Floating point number")
def get_prep_value(self, value):
value = super(FloatField, self).get_prep_value(value)
if value is None:
return None
return float(value)
def get_internal_type(self):
return "FloatField"
def to_python(self, value):
if value is None:
return value
try:
return float(value)
except (TypeError, ValueError):
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def formfield(self, **kwargs):
defaults = {'form_class': forms.FloatField}
defaults.update(kwargs)
return super(FloatField, self).formfield(**defaults)
class IntegerField(Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be an integer."),
}
description = _("Integer")
def check(self, **kwargs):
errors = super(IntegerField, self).check(**kwargs)
errors.extend(self._check_max_length_warning())
return errors
def _check_max_length_warning(self):
if self.max_length is not None:
return [
checks.Warning(
"'max_length' is ignored when used with IntegerField",
hint="Remove 'max_length' from field",
obj=self,
id='fields.W122',
)
]
return []
@cached_property
def validators(self):
# These validators can't be added at field initialization time since
# they're based on values retrieved from `connection`.
range_validators = []
internal_type = self.get_internal_type()
min_value, max_value = connection.ops.integer_field_range(internal_type)
if min_value is not None:
range_validators.append(validators.MinValueValidator(min_value))
if max_value is not None:
range_validators.append(validators.MaxValueValidator(max_value))
return super(IntegerField, self).validators + range_validators
def get_prep_value(self, value):
value = super(IntegerField, self).get_prep_value(value)
if value is None:
return None
return int(value)
def get_prep_lookup(self, lookup_type, value):
if ((lookup_type == 'gte' or lookup_type == 'lt')
and isinstance(value, float)):
value = math.ceil(value)
return super(IntegerField, self).get_prep_lookup(lookup_type, value)
def get_internal_type(self):
return "IntegerField"
def to_python(self, value):
if value is None:
return value
try:
return int(value)
except (TypeError, ValueError):
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def formfield(self, **kwargs):
defaults = {'form_class': forms.IntegerField}
defaults.update(kwargs)
return super(IntegerField, self).formfield(**defaults)
class BigIntegerField(IntegerField):
empty_strings_allowed = False
description = _("Big (8 byte) integer")
MAX_BIGINT = 9223372036854775807
def get_internal_type(self):
return "BigIntegerField"
def formfield(self, **kwargs):
defaults = {'min_value': -BigIntegerField.MAX_BIGINT - 1,
'max_value': BigIntegerField.MAX_BIGINT}
defaults.update(kwargs)
return super(BigIntegerField, self).formfield(**defaults)
class IPAddressField(Field):
empty_strings_allowed = False
description = _("IPv4 address")
system_check_deprecated_details = {
'msg': (
'IPAddressField has been deprecated. Support for it (except in '
'historical migrations) will be removed in Django 1.9.'
),
'hint': 'Use GenericIPAddressField instead.',
'id': 'fields.W900',
}
def __init__(self, *args, **kwargs):
kwargs['max_length'] = 15
super(IPAddressField, self).__init__(*args, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(IPAddressField, self).deconstruct()
del kwargs['max_length']
return name, path, args, kwargs
def get_prep_value(self, value):
value = super(IPAddressField, self).get_prep_value(value)
if value is None:
return None
return six.text_type(value)
def get_internal_type(self):
return "IPAddressField"
def formfield(self, **kwargs):
defaults = {'form_class': forms.IPAddressField}
defaults.update(kwargs)
return super(IPAddressField, self).formfield(**defaults)
class GenericIPAddressField(Field):
empty_strings_allowed = False
description = _("IP address")
default_error_messages = {}
def __init__(self, verbose_name=None, name=None, protocol='both',
unpack_ipv4=False, *args, **kwargs):
self.unpack_ipv4 = unpack_ipv4
self.protocol = protocol
self.default_validators, invalid_error_message = \
validators.ip_address_validators(protocol, unpack_ipv4)
self.default_error_messages['invalid'] = invalid_error_message
kwargs['max_length'] = 39
super(GenericIPAddressField, self).__init__(verbose_name, name, *args,
**kwargs)
def check(self, **kwargs):
errors = super(GenericIPAddressField, self).check(**kwargs)
errors.extend(self._check_blank_and_null_values(**kwargs))
return errors
def _check_blank_and_null_values(self, **kwargs):
if not getattr(self, 'null', False) and getattr(self, 'blank', False):
return [
checks.Error(
('GenericIPAddressFields cannot have blank=True if null=False, '
'as blank values are stored as nulls.'),
hint=None,
obj=self,
id='fields.E150',
)
]
return []
def deconstruct(self):
name, path, args, kwargs = super(GenericIPAddressField, self).deconstruct()
if self.unpack_ipv4 is not False:
kwargs['unpack_ipv4'] = self.unpack_ipv4
if self.protocol != "both":
kwargs['protocol'] = self.protocol
if kwargs.get("max_length", None) == 39:
del kwargs['max_length']
return name, path, args, kwargs
def get_internal_type(self):
return "GenericIPAddressField"
def to_python(self, value):
if value and ':' in value:
return clean_ipv6_address(value,
self.unpack_ipv4, self.error_messages['invalid'])
return value
def get_db_prep_value(self, value, connection, prepared=False):
if not prepared:
value = self.get_prep_value(value)
return connection.ops.value_to_db_ipaddress(value)
def get_prep_value(self, value):
value = super(GenericIPAddressField, self).get_prep_value(value)
if value is None:
return None
if value and ':' in value:
try:
return clean_ipv6_address(value, self.unpack_ipv4)
except exceptions.ValidationError:
pass
return six.text_type(value)
def formfield(self, **kwargs):
defaults = {
'protocol': self.protocol,
'form_class': forms.GenericIPAddressField,
}
defaults.update(kwargs)
return super(GenericIPAddressField, self).formfield(**defaults)
class NullBooleanField(Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value must be either None, True or False."),
}
description = _("Boolean (Either True, False or None)")
def __init__(self, *args, **kwargs):
kwargs['null'] = True
kwargs['blank'] = True
super(NullBooleanField, self).__init__(*args, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(NullBooleanField, self).deconstruct()
del kwargs['null']
del kwargs['blank']
return name, path, args, kwargs
def get_internal_type(self):
return "NullBooleanField"
def to_python(self, value):
if value is None:
return None
if value in (True, False):
return bool(value)
if value in ('None',):
return None
if value in ('t', 'True', '1'):
return True
if value in ('f', 'False', '0'):
return False
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def get_prep_lookup(self, lookup_type, value):
# Special-case handling for filters coming from a Web request (e.g. the
# admin interface). Only works for scalar values (not lists). If you're
# passing in a list, you might as well make things the right type when
# constructing the list.
if value in ('1', '0'):
value = bool(int(value))
return super(NullBooleanField, self).get_prep_lookup(lookup_type,
value)
def get_prep_value(self, value):
value = super(NullBooleanField, self).get_prep_value(value)
if value is None:
return None
return bool(value)
def formfield(self, **kwargs):
defaults = {
'form_class': forms.NullBooleanField,
'required': not self.blank,
'label': capfirst(self.verbose_name),
'help_text': self.help_text}
defaults.update(kwargs)
return super(NullBooleanField, self).formfield(**defaults)
class PositiveIntegerField(IntegerField):
description = _("Positive integer")
def get_internal_type(self):
return "PositiveIntegerField"
def formfield(self, **kwargs):
defaults = {'min_value': 0}
defaults.update(kwargs)
return super(PositiveIntegerField, self).formfield(**defaults)
class PositiveSmallIntegerField(IntegerField):
description = _("Positive small integer")
def get_internal_type(self):
return "PositiveSmallIntegerField"
def formfield(self, **kwargs):
defaults = {'min_value': 0}
defaults.update(kwargs)
return super(PositiveSmallIntegerField, self).formfield(**defaults)
class SlugField(CharField):
default_validators = [validators.validate_slug]
description = _("Slug (up to %(max_length)s)")
def __init__(self, *args, **kwargs):
kwargs['max_length'] = kwargs.get('max_length', 50)
# Set db_index=True unless it's been set manually.
if 'db_index' not in kwargs:
kwargs['db_index'] = True
super(SlugField, self).__init__(*args, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(SlugField, self).deconstruct()
if kwargs.get("max_length", None) == 50:
del kwargs['max_length']
if self.db_index is False:
kwargs['db_index'] = False
else:
del kwargs['db_index']
return name, path, args, kwargs
def get_internal_type(self):
return "SlugField"
def formfield(self, **kwargs):
defaults = {'form_class': forms.SlugField}
defaults.update(kwargs)
return super(SlugField, self).formfield(**defaults)
class SmallIntegerField(IntegerField):
description = _("Small integer")
def get_internal_type(self):
return "SmallIntegerField"
class TextField(Field):
description = _("Text")
def get_internal_type(self):
return "TextField"
def get_prep_value(self, value):
value = super(TextField, self).get_prep_value(value)
if isinstance(value, six.string_types) or value is None:
return value
return smart_text(value)
def formfield(self, **kwargs):
# Passing max_length to forms.CharField means that the value's length
# will be validated twice. This is considered acceptable since we want
# the value in the form field (to pass into widget for example).
defaults = {'max_length': self.max_length, 'widget': forms.Textarea}
defaults.update(kwargs)
return super(TextField, self).formfield(**defaults)
class TimeField(DateTimeCheckMixin, Field):
empty_strings_allowed = False
default_error_messages = {
'invalid': _("'%(value)s' value has an invalid format. It must be in "
"HH:MM[:ss[.uuuuuu]] format."),
'invalid_time': _("'%(value)s' value has the correct format "
"(HH:MM[:ss[.uuuuuu]]) but it is an invalid time."),
}
description = _("Time")
def __init__(self, verbose_name=None, name=None, auto_now=False,
auto_now_add=False, **kwargs):
self.auto_now, self.auto_now_add = auto_now, auto_now_add
if auto_now or auto_now_add:
kwargs['editable'] = False
kwargs['blank'] = True
super(TimeField, self).__init__(verbose_name, name, **kwargs)
def _check_fix_default_value(self):
"""
Adds a warning to the checks framework stating, that using an actual
time or datetime value is probably wrong; it's only being evaluated on
server start-up.
For details see ticket #21905
"""
if not self.has_default():
return []
now = timezone.now()
if not timezone.is_naive(now):
now = timezone.make_naive(now, timezone.utc)
value = self.default
if isinstance(value, datetime.datetime):
second_offset = datetime.timedelta(seconds=10)
lower = now - second_offset
upper = now + second_offset
if timezone.is_aware(value):
value = timezone.make_naive(value, timezone.utc)
elif isinstance(value, datetime.time):
second_offset = datetime.timedelta(seconds=10)
lower = now - second_offset
upper = now + second_offset
value = datetime.datetime.combine(now.date(), value)
if timezone.is_aware(value):
value = timezone.make_naive(value, timezone.utc).time()
else:
# No explicit time / datetime value -- no checks necessary
return []
if lower <= value <= upper:
return [
checks.Warning(
'Fixed default value provided.',
hint='It seems you set a fixed date / time / datetime '
'value as default for this field. This may not be '
'what you want. If you want to have the current date '
'as default, use `django.utils.timezone.now`',
obj=self,
id='fields.W161',
)
]
return []
def deconstruct(self):
name, path, args, kwargs = super(TimeField, self).deconstruct()
if self.auto_now is not False:
kwargs["auto_now"] = self.auto_now
if self.auto_now_add is not False:
kwargs["auto_now_add"] = self.auto_now_add
if self.auto_now or self.auto_now_add:
del kwargs['blank']
del kwargs['editable']
return name, path, args, kwargs
def get_internal_type(self):
return "TimeField"
def to_python(self, value):
if value is None:
return None
if isinstance(value, datetime.time):
return value
if isinstance(value, datetime.datetime):
# Not usually a good idea to pass in a datetime here (it loses
# information), but this can be a side-effect of interacting with a
# database backend (e.g. Oracle), so we'll be accommodating.
return value.time()
try:
parsed = parse_time(value)
if parsed is not None:
return parsed
except ValueError:
raise exceptions.ValidationError(
self.error_messages['invalid_time'],
code='invalid_time',
params={'value': value},
)
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
def pre_save(self, model_instance, add):
if self.auto_now or (self.auto_now_add and add):
value = datetime.datetime.now().time()
setattr(model_instance, self.attname, value)
return value
else:
return super(TimeField, self).pre_save(model_instance, add)
def get_prep_value(self, value):
value = super(TimeField, self).get_prep_value(value)
return self.to_python(value)
def get_db_prep_value(self, value, connection, prepared=False):
# Casts times into the format expected by the backend
if not prepared:
value = self.get_prep_value(value)
return connection.ops.value_to_db_time(value)
def value_to_string(self, obj):
val = self._get_val_from_obj(obj)
return '' if val is None else val.isoformat()
def formfield(self, **kwargs):
defaults = {'form_class': forms.TimeField}
defaults.update(kwargs)
return super(TimeField, self).formfield(**defaults)
class URLField(CharField):
default_validators = [validators.URLValidator()]
description = _("URL")
def __init__(self, verbose_name=None, name=None, **kwargs):
kwargs['max_length'] = kwargs.get('max_length', 200)
super(URLField, self).__init__(verbose_name, name, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(URLField, self).deconstruct()
if kwargs.get("max_length", None) == 200:
del kwargs['max_length']
return name, path, args, kwargs
def formfield(self, **kwargs):
# As with CharField, this will cause URL validation to be performed
# twice.
defaults = {
'form_class': forms.URLField,
}
defaults.update(kwargs)
return super(URLField, self).formfield(**defaults)
class BinaryField(Field):
description = _("Raw binary data")
empty_values = [None, b'']
def __init__(self, *args, **kwargs):
kwargs['editable'] = False
super(BinaryField, self).__init__(*args, **kwargs)
if self.max_length is not None:
self.validators.append(validators.MaxLengthValidator(self.max_length))
def deconstruct(self):
name, path, args, kwargs = super(BinaryField, self).deconstruct()
del kwargs['editable']
return name, path, args, kwargs
def get_internal_type(self):
return "BinaryField"
def get_default(self):
if self.has_default() and not callable(self.default):
return self.default
default = super(BinaryField, self).get_default()
if default == '':
return b''
return default
def get_db_prep_value(self, value, connection, prepared=False):
value = super(BinaryField, self).get_db_prep_value(value, connection, prepared)
if value is not None:
return connection.Database.Binary(value)
return value
def value_to_string(self, obj):
"""Binary data is serialized as base64"""
return b64encode(force_bytes(self._get_val_from_obj(obj))).decode('ascii')
def to_python(self, value):
# If it's a string, it should be base64-encoded data
if isinstance(value, six.text_type):
return six.memoryview(b64decode(force_bytes(value)))
return value
class UUIDField(Field):
default_error_messages = {
'invalid': _("'%(value)s' is not a valid UUID."),
}
description = 'Universally unique identifier'
empty_strings_allowed = False
def __init__(self, verbose_name=None, **kwargs):
kwargs['max_length'] = 32
super(UUIDField, self).__init__(verbose_name, **kwargs)
def deconstruct(self):
name, path, args, kwargs = super(UUIDField, self).deconstruct()
del kwargs['max_length']
return name, path, args, kwargs
def get_internal_type(self):
return "UUIDField"
def get_db_prep_value(self, value, connection, prepared=False):
if isinstance(value, six.string_types):
value = uuid.UUID(value.replace('-', ''))
if isinstance(value, uuid.UUID):
if connection.features.has_native_uuid_field:
return value
return value.hex
return value
def to_python(self, value):
if value and not isinstance(value, uuid.UUID):
try:
return uuid.UUID(value)
except ValueError:
raise exceptions.ValidationError(
self.error_messages['invalid'],
code='invalid',
params={'value': value},
)
return value
def formfield(self, **kwargs):
defaults = {
'form_class': forms.UUIDField,
}
defaults.update(kwargs)
return super(UUIDField, self).formfield(**defaults)
| [
"multics_luo@163.com"
] | multics_luo@163.com |
dede7897aa89f5106650a0cc96628e0b6108d67e | 9ee537601cbca4be70870d60910ce808da4f766e | /PROJECT-WORKING VERSION-issue.py | bcf71964dd23b6d934705ca27ceb5766335b8449 | [] | no_license | sanjanab003/Library-Management | 063b98e98d50bbd16e358329b7683707561965c6 | eafcbc7814a7933c6243ef6cec02b803c4810a2d | refs/heads/master | 2020-08-01T01:51:46.160900 | 2019-10-20T04:48:08 | 2019-10-20T04:48:08 | 210,818,939 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 3,914 | py | c=0
areMem=()
c1=0
t=()
u=""
p=""
q=""
b=0
a=0
user=""
pas=""
z=0
t1=()
import mysql.connector
mydb = mysql.connector.connect(
host="localhost",
user="root",
passwd="0025",
database="library"
)
mycursor = mydb.cursor()
def regName():
print("Please register")
global u
global p
u=input("Enter your name: ")
p=input("Enter a password: ")
# function to register
def reg():
add_member = ("INSERT INTO member "
"(Name, M_Pass) "
"VALUES (%s, %s)")
data_member = (u,p)
mycursor.execute(add_member, data_member)
mydb.commit()
# function to loop through members' table
def loopMem():
mycursor.execute("select * from member")
global c
global t1
row = mycursor.fetchone()
while row is not None:
if(row[0]==pas and row[1]==user):
c=c+1
t1=t1+(row,)
row=mycursor.fetchone()
else:
row = mycursor.fetchone()
def areMem():
global areMem
areMem=input("Are you a already a member of Sanchansu(Type in 'y' for yes and 'n' for no): ")
def ask():
global user
global pas
user=input("Enter username: ")
pas=input("Enter password: ")
# function to loop through the books
def loopBook(d):
mycursor.execute("select * from books")
global c1
row=mycursor.fetchone()
while row is not None:
if(row[1]==d):
global t
t=t+(row,)
c1=c1+1
row=mycursor.fetchone()
else:
row=mycursor.fetchone()
# function to ask for user's choice
def inp():
global q
q=input("Enter the name of the book you want to borrow: ")
# function to issue book
def iss(e,f):
op= """ update books set Qty = %s where Book_Name = %s """
op3= """ update member set Books_borrowed = %s where M_Pass = %s"""
dat=(e,f)
dat3=(f,pas)
mycursor.execute(op,dat)
mycursor.execute(op3,dat3)
if(e==0):
op2= """ update books set Status = %s where Book_Name= %s """
dat2= ("Not available", f)
mycursor.execute(op2,dat2)
else:
pass;
mydb.commit()
print("Book borrowed successfully")
def tup():
global a
global b
for x in t:
a=x[1]
b=x[4]-1
def tup1():
global z
for p in t1:
z=p[2]
def issue(r):
inp()
loopBook(q)
tup()
if(c1>0):
print(t)
if(b+1>0):
q2=input("Do you want to borrow this book: ")
if(q2=='y'):
if(r=="NULL"):
iss(b,a)
else:
print("Please return book before borrowing")
else:
pass;
else:
print("Sorry Book currently not available")
else:
print("Sorry no matching results")
# checking for membership
areMem()
if(areMem=="y"):
ask()
loopMem()
tup1()
if(c==1):
print("Welcome,", user)
print("Books borrowed by you: ", z)
print("Books to be returned: ", z)
issue(z)
else:
print("Incorrect password or username")
elif(areMem=="n"):
regName()
loopMem()
if(c==1):
print("Username or password is already taken")
else:
reg()
print("You are now a member of Sanchansu.")
print()
issue()
else:
print("Please give a valid answer")
| [
"noreply@github.com"
] | noreply@github.com |
814949dee15c29811a0f3cb2131cabed7e0f87f0 | 52b17383556c615f6e9ea8888efbfc3c0bfbe726 | /tests/__init__.py | 2aac0df30f37c2d7f9d7ec10bada80f300ecf814 | [
"Apache-2.0"
] | permissive | nerdfiles/bloom-filter | 725f164bd5b99339f128d34f6b355f4efdbdc0f0 | affd787cec3804fb806146eb464f77f37ed5b85e | refs/heads/master | 2021-09-15T03:16:02.775833 | 2018-05-25T00:02:02 | 2018-05-25T00:02:02 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 480 | py | #-----------------------------------------------------------------------------#
# __init__.py #
# #
# Copyright (c) 2017-2018, Rajiv Bakulesh Shah, original author. #
# All rights reserved. #
#-----------------------------------------------------------------------------#
| [
"noreply@github.com"
] | noreply@github.com |
7728d77ed32221009bb1df33c077dfc8ef3d3b5f | 8dcd3ee098b4f5b80879c37a62292f42f6b2ae17 | /venv/Lib/site-packages/pandas/tests/frame/test_api.py | 1cfbd5961177ae6b74b0110ac71414299c5047fe | [] | no_license | GregVargas1999/InfinityAreaInfo | 53fdfefc11c4af8f5d2b8f511f7461d11a3f7533 | 2e4a7c6a2424514ca0ec58c9153eb08dc8e09a4a | refs/heads/master | 2022-12-01T20:26:05.388878 | 2020-08-11T18:37:05 | 2020-08-11T18:37:05 | 286,821,452 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 18,965 | py | import datetime
import pydoc
from copy import deepcopy
import numpy as np
import pandas as pd
import pandas._testing as tm
import pytest
from pandas import Categorical, DataFrame, Series, compat, date_range, timedelta_range
from pandas.compat import PY37
from pandas.util._test_decorators import async_mark
class TestDataFrameMisc:
def test_copy_index_name_checking(self, float_frame):
# don't want to be able to modify the index stored elsewhere after
# making a copy
for attr in ("index", "columns"):
ind = getattr(float_frame, attr)
ind.name = None
cp = float_frame.copy()
getattr(cp, attr).name = "foo"
assert getattr(float_frame, attr).name is None
def test_getitem_pop_assign_name(self, float_frame):
s = float_frame["A"]
assert s.name == "A"
s = float_frame.pop("A")
assert s.name == "A"
s = float_frame.loc[:, "B"]
assert s.name == "B"
s2 = s.loc[:]
assert s2.name == "B"
def test_get_value(self, float_frame):
for idx in float_frame.index:
for col in float_frame.columns:
result = float_frame._get_value(idx, col)
expected = float_frame[col][idx]
tm.assert_almost_equal(result, expected)
def test_add_prefix_suffix(self, float_frame):
with_prefix = float_frame.add_prefix("foo#")
expected = pd.Index(["foo#{c}".format(c=c) for c in float_frame.columns])
tm.assert_index_equal(with_prefix.columns, expected)
with_suffix = float_frame.add_suffix("#foo")
expected = pd.Index(["{c}#foo".format(c=c) for c in float_frame.columns])
tm.assert_index_equal(with_suffix.columns, expected)
with_pct_prefix = float_frame.add_prefix("%")
expected = pd.Index(["%{c}".format(c=c) for c in float_frame.columns])
tm.assert_index_equal(with_pct_prefix.columns, expected)
with_pct_suffix = float_frame.add_suffix("%")
expected = pd.Index(["{c}%".format(c=c) for c in float_frame.columns])
tm.assert_index_equal(with_pct_suffix.columns, expected)
def test_get_axis(self, float_frame):
f = float_frame
assert f._get_axis_number(0) == 0
assert f._get_axis_number(1) == 1
assert f._get_axis_number("index") == 0
assert f._get_axis_number("rows") == 0
assert f._get_axis_number("columns") == 1
assert f._get_axis_name(0) == "index"
assert f._get_axis_name(1) == "columns"
assert f._get_axis_name("index") == "index"
assert f._get_axis_name("rows") == "index"
assert f._get_axis_name("columns") == "columns"
assert f._get_axis(0) is f.index
assert f._get_axis(1) is f.columns
with pytest.raises(ValueError, match="No axis named"):
f._get_axis_number(2)
with pytest.raises(ValueError, match="No axis.*foo"):
f._get_axis_name("foo")
with pytest.raises(ValueError, match="No axis.*None"):
f._get_axis_name(None)
with pytest.raises(ValueError, match="No axis named"):
f._get_axis_number(None)
def test_keys(self, float_frame):
getkeys = float_frame.keys
assert getkeys() is float_frame.columns
def test_column_contains_raises(self, float_frame):
with pytest.raises(TypeError, match="unhashable type: 'Index'"):
float_frame.columns in float_frame
def test_tab_completion(self):
# DataFrame whose columns are identifiers shall have them in __dir__.
df = pd.DataFrame([list("abcd"), list("efgh")], columns=list("ABCD"))
for key in list("ABCD"):
assert key in dir(df)
assert isinstance(df.__getitem__("A"), pd.Series)
# DataFrame whose first-level columns are identifiers shall have
# them in __dir__.
df = pd.DataFrame(
[list("abcd"), list("efgh")],
columns=pd.MultiIndex.from_tuples(list(zip("ABCD", "EFGH"))),
)
for key in list("ABCD"):
assert key in dir(df)
for key in list("EFGH"):
assert key not in dir(df)
assert isinstance(df.__getitem__("A"), pd.DataFrame)
def test_not_hashable(self):
empty_frame = DataFrame()
df = DataFrame([1])
msg = "'DataFrame' objects are mutable, thus they cannot be hashed"
with pytest.raises(TypeError, match=msg):
hash(df)
with pytest.raises(TypeError, match=msg):
hash(empty_frame)
def test_new_empty_index(self):
df1 = DataFrame(np.random.randn(0, 3))
df2 = DataFrame(np.random.randn(0, 3))
df1.index.name = "foo"
assert df2.index.name is None
def test_array_interface(self, float_frame):
with np.errstate(all="ignore"):
result = np.sqrt(float_frame)
assert isinstance(result, type(float_frame))
assert result.index is float_frame.index
assert result.columns is float_frame.columns
tm.assert_frame_equal(result, float_frame.apply(np.sqrt))
def test_get_agg_axis(self, float_frame):
cols = float_frame._get_agg_axis(0)
assert cols is float_frame.columns
idx = float_frame._get_agg_axis(1)
assert idx is float_frame.index
msg = r"Axis must be 0 or 1 \(got 2\)"
with pytest.raises(ValueError, match=msg):
float_frame._get_agg_axis(2)
def test_nonzero(self, float_frame, float_string_frame):
empty_frame = DataFrame()
assert empty_frame.empty
assert not float_frame.empty
assert not float_string_frame.empty
# corner case
df = DataFrame({"A": [1.0, 2.0, 3.0], "B": ["a", "b", "c"]}, index=np.arange(3))
del df["A"]
assert not df.empty
def test_iteritems(self):
df = DataFrame([[1, 2, 3], [4, 5, 6]], columns=["a", "a", "b"])
for k, v in df.items():
assert isinstance(v, DataFrame._constructor_sliced)
def test_items(self):
# GH 17213, GH 13918
cols = ["a", "b", "c"]
df = DataFrame([[1, 2, 3], [4, 5, 6]], columns=cols)
for c, (k, v) in zip(cols, df.items()):
assert c == k
assert isinstance(v, Series)
assert (df[k] == v).all()
def test_iter(self, float_frame):
assert tm.equalContents(list(float_frame), float_frame.columns)
def test_iterrows(self, float_frame, float_string_frame):
for k, v in float_frame.iterrows():
exp = float_frame.loc[k]
tm.assert_series_equal(v, exp)
for k, v in float_string_frame.iterrows():
exp = float_string_frame.loc[k]
tm.assert_series_equal(v, exp)
def test_iterrows_iso8601(self):
# GH 19671
s = DataFrame(
{
"non_iso8601": ["M1701", "M1802", "M1903", "M2004"],
"iso8601": date_range("2000-01-01", periods=4, freq="M"),
}
)
for k, v in s.iterrows():
exp = s.loc[k]
tm.assert_series_equal(v, exp)
def test_iterrows_corner(self):
# gh-12222
df = DataFrame(
{
"a": [datetime.datetime(2015, 1, 1)],
"b": [None],
"c": [None],
"d": [""],
"e": [[]],
"f": [set()],
"g": [{}],
}
)
expected = Series(
[datetime.datetime(2015, 1, 1), None, None, "", [], set(), {}],
index=list("abcdefg"),
name=0,
dtype="object",
)
_, result = next(df.iterrows())
tm.assert_series_equal(result, expected)
def test_itertuples(self, float_frame):
for i, tup in enumerate(float_frame.itertuples()):
s = DataFrame._constructor_sliced(tup[1:])
s.name = tup[0]
expected = float_frame.iloc[i, :].reset_index(drop=True)
tm.assert_series_equal(s, expected)
df = DataFrame(
{"floats": np.random.randn(5), "ints": range(5)}, columns=["floats", "ints"]
)
for tup in df.itertuples(index=False):
assert isinstance(tup[1], int)
df = DataFrame(data={"a": [1, 2, 3], "b": [4, 5, 6]})
dfaa = df[["a", "a"]]
assert list(dfaa.itertuples()) == [(0, 1, 1), (1, 2, 2), (2, 3, 3)]
# repr with int on 32-bit/windows
if not (compat.is_platform_windows() or compat.is_platform_32bit()):
assert (
repr(list(df.itertuples(name=None)))
== "[(0, 1, 4), (1, 2, 5), (2, 3, 6)]"
)
tup = next(df.itertuples(name="TestName"))
assert tup._fields == ("Index", "a", "b")
assert (tup.Index, tup.a, tup.b) == tup
assert type(tup).__name__ == "TestName"
df.columns = ["def", "return"]
tup2 = next(df.itertuples(name="TestName"))
assert tup2 == (0, 1, 4)
assert tup2._fields == ("Index", "_1", "_2")
df3 = DataFrame({"f" + str(i): [i] for i in range(1024)})
# will raise SyntaxError if trying to create namedtuple
tup3 = next(df3.itertuples())
assert isinstance(tup3, tuple)
if PY37:
assert hasattr(tup3, "_fields")
else:
assert not hasattr(tup3, "_fields")
# GH 28282
df_254_columns = DataFrame([{f"foo_{i}": f"bar_{i}" for i in range(254)}])
result_254_columns = next(df_254_columns.itertuples(index=False))
assert isinstance(result_254_columns, tuple)
assert hasattr(result_254_columns, "_fields")
df_255_columns = DataFrame([{f"foo_{i}": f"bar_{i}" for i in range(255)}])
result_255_columns = next(df_255_columns.itertuples(index=False))
assert isinstance(result_255_columns, tuple)
# Dataframes with >=255 columns will fallback to regular tuples on python < 3.7
if PY37:
assert hasattr(result_255_columns, "_fields")
else:
assert not hasattr(result_255_columns, "_fields")
def test_sequence_like_with_categorical(self):
# GH 7839
# make sure can iterate
df = DataFrame(
{"id": [1, 2, 3, 4, 5, 6], "raw_grade": ["a", "b", "b", "a", "a", "e"]}
)
df["grade"] = Categorical(df["raw_grade"])
# basic sequencing testing
result = list(df.grade.values)
expected = np.array(df.grade.values).tolist()
tm.assert_almost_equal(result, expected)
# iteration
for t in df.itertuples(index=False):
str(t)
for row, s in df.iterrows():
str(s)
for c, col in df.items():
str(s)
def test_len(self, float_frame):
assert len(float_frame) == len(float_frame.index)
def test_values_mixed_dtypes(self, float_frame, float_string_frame):
frame = float_frame
arr = frame.values
frame_cols = frame.columns
for i, row in enumerate(arr):
for j, value in enumerate(row):
col = frame_cols[j]
if np.isnan(value):
assert np.isnan(frame[col][i])
else:
assert value == frame[col][i]
# mixed type
arr = float_string_frame[["foo", "A"]].values
assert arr[0, 0] == "bar"
df = DataFrame({"complex": [1j, 2j, 3j], "real": [1, 2, 3]})
arr = df.values
assert arr[0, 0] == 1j
# single block corner case
arr = float_frame[["A", "B"]].values
expected = float_frame.reindex(columns=["A", "B"]).values
tm.assert_almost_equal(arr, expected)
def test_to_numpy(self):
df = pd.DataFrame({"A": [1, 2], "B": [3, 4.5]})
expected = np.array([[1, 3], [2, 4.5]])
result = df.to_numpy()
tm.assert_numpy_array_equal(result, expected)
def test_to_numpy_dtype(self):
df = pd.DataFrame({"A": [1, 2], "B": [3, 4.5]})
expected = np.array([[1, 3], [2, 4]], dtype="int64")
result = df.to_numpy(dtype="int64")
tm.assert_numpy_array_equal(result, expected)
def test_to_numpy_copy(self):
arr = np.random.randn(4, 3)
df = pd.DataFrame(arr)
assert df.values.base is arr
assert df.to_numpy(copy=False).base is arr
assert df.to_numpy(copy=True).base is None
def test_transpose(self, float_frame):
frame = float_frame
dft = frame.T
for idx, series in dft.items():
for col, value in series.items():
if np.isnan(value):
assert np.isnan(frame[col][idx])
else:
assert value == frame[col][idx]
# mixed type
index, data = tm.getMixedTypeDict()
mixed = DataFrame(data, index=index)
mixed_T = mixed.T
for col, s in mixed_T.items():
assert s.dtype == np.object_
def test_swapaxes(self):
df = DataFrame(np.random.randn(10, 5))
tm.assert_frame_equal(df.T, df.swapaxes(0, 1))
tm.assert_frame_equal(df.T, df.swapaxes(1, 0))
tm.assert_frame_equal(df, df.swapaxes(0, 0))
msg = (
"No axis named 2 for object type"
r" <class 'pandas.core(.sparse)?.frame.(Sparse)?DataFrame'>"
)
with pytest.raises(ValueError, match=msg):
df.swapaxes(2, 5)
def test_axis_aliases(self, float_frame):
f = float_frame
# reg name
expected = f.sum(axis=0)
result = f.sum(axis="index")
tm.assert_series_equal(result, expected)
expected = f.sum(axis=1)
result = f.sum(axis="columns")
tm.assert_series_equal(result, expected)
def test_class_axis(self):
# GH 18147
# no exception and no empty docstring
assert pydoc.getdoc(DataFrame.index)
assert pydoc.getdoc(DataFrame.columns)
def test_more_values(self, float_string_frame):
values = float_string_frame.values
assert values.shape[1] == len(float_string_frame.columns)
def test_repr_with_mi_nat(self, float_string_frame):
df = DataFrame(
{"X": [1, 2]}, index=[[pd.NaT, pd.Timestamp("20130101")], ["a", "b"]]
)
result = repr(df)
expected = " X\nNaT a 1\n2013-01-01 b 2"
assert result == expected
def test_items_names(self, float_string_frame):
for k, v in float_string_frame.items():
assert v.name == k
def test_series_put_names(self, float_string_frame):
series = float_string_frame._series
for k, v in series.items():
assert v.name == k
def test_empty_nonzero(self):
df = DataFrame([1, 2, 3])
assert not df.empty
df = DataFrame(index=[1], columns=[1])
assert not df.empty
df = DataFrame(index=["a", "b"], columns=["c", "d"]).dropna()
assert df.empty
assert df.T.empty
empty_frames = [
DataFrame(),
DataFrame(index=[1]),
DataFrame(columns=[1]),
DataFrame({1: []}),
]
for df in empty_frames:
assert df.empty
assert df.T.empty
def test_with_datetimelikes(self):
df = DataFrame(
{
"A": date_range("20130101", periods=10),
"B": timedelta_range("1 day", periods=10),
}
)
t = df.T
result = t.dtypes.value_counts()
expected = Series({np.dtype("object"): 10})
tm.assert_series_equal(result, expected)
def test_values(self, float_frame):
float_frame.values[:, 0] = 5.0
assert (float_frame.values[:, 0] == 5).all()
def test_deepcopy(self, float_frame):
cp = deepcopy(float_frame)
series = cp["A"]
series[:] = 10
for idx, value in series.items():
assert float_frame["A"][idx] != value
def test_transpose_get_view(self, float_frame):
dft = float_frame.T
dft.values[:, 5:10] = 5
assert (float_frame.values[5:10] == 5).all()
def test_inplace_return_self(self):
# GH 1893
data = DataFrame(
{"a": ["foo", "bar", "baz", "qux"], "b": [0, 0, 1, 1], "c": [1, 2, 3, 4]}
)
def _check_f(base, f):
result = f(base)
assert result is None
# -----DataFrame-----
# set_index
f = lambda x: x.set_index("a", inplace=True)
_check_f(data.copy(), f)
# reset_index
f = lambda x: x.reset_index(inplace=True)
_check_f(data.set_index("a"), f)
# drop_duplicates
f = lambda x: x.drop_duplicates(inplace=True)
_check_f(data.copy(), f)
# sort
f = lambda x: x.sort_values("b", inplace=True)
_check_f(data.copy(), f)
# sort_index
f = lambda x: x.sort_index(inplace=True)
_check_f(data.copy(), f)
# fillna
f = lambda x: x.fillna(0, inplace=True)
_check_f(data.copy(), f)
# replace
f = lambda x: x.replace(1, 0, inplace=True)
_check_f(data.copy(), f)
# rename
f = lambda x: x.rename({1: "foo"}, inplace=True)
_check_f(data.copy(), f)
# -----Series-----
d = data.copy()["c"]
# reset_index
f = lambda x: x.reset_index(inplace=True, drop=True)
_check_f(data.set_index("a")["c"], f)
# fillna
f = lambda x: x.fillna(0, inplace=True)
_check_f(d.copy(), f)
# replace
f = lambda x: x.replace(1, 0, inplace=True)
_check_f(d.copy(), f)
# rename
f = lambda x: x.rename({1: "foo"}, inplace=True)
_check_f(d.copy(), f)
@async_mark()
async def test_tab_complete_warning(self, ip):
# GH 16409
pytest.importorskip("IPython", minversion="6.0.0")
from IPython.core.completer import provisionalcompleter
code = "import pandas as pd; df = pd.DataFrame()"
await ip.run_code(code)
# TODO: remove it when Ipython updates
# GH 33567, jedi version raises Deprecation warning in Ipython
import jedi
if jedi.__version__ < "0.17.0":
warning = tm.assert_produces_warning(None)
else:
warning = tm.assert_produces_warning(
DeprecationWarning, check_stacklevel=False
)
with warning:
with provisionalcompleter("ignore"):
list(ip.Completer.completions("df.", 1))
def test_attrs(self):
df = pd.DataFrame({"A": [2, 3]})
assert df.attrs == {}
df.attrs["version"] = 1
result = df.rename(columns=str)
assert result.attrs == {"version": 1}
| [
"44142880+GregVargas1999@users.noreply.github.com"
] | 44142880+GregVargas1999@users.noreply.github.com |
94d95349f260143004954f60708415d49af4c291 | 741d2868877fbd237a966e04899942224e36ae76 | /QT_Mouse_House/Qt_Exp_Controls.py | dba439c0336c290c9bc35c28e9d3e221e52bcfe6 | [] | no_license | TiangeLi/ArduinoCntrl | 5e7b9357a4b304e712b1046ae6d13cda05c4fc01 | f0ec2eef91d145f4807928e020857c62aaaf159b | refs/heads/master | 2021-05-04T09:41:44.942708 | 2018-03-28T18:15:44 | 2018-03-28T18:15:44 | 61,747,298 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 37,091 | py | # coding=utf-8
"""Qt Widgets for Experiment Config and Control"""
import os
from Names import *
from Misc_Classes import *
from Misc_Functions import *
from Custom_Qt_Tools import *
from copy import deepcopy
from Dirs_Settings import ArdDataContainer
from Custom_Qt_Widgets import GUI_LiveScrollingGraph, GUI_LJDataReport, GUI_StatusBars
import PyQt4.QtGui as qg
import PyQt4.QtCore as qc
class GUI_ExpControls(qg.QWidget):
"""Settings and Buttons for Configuring the Experiment"""
def __init__(self, dirs, gui_progbar):
qg.QWidget.__init__(self)
self.dirs = dirs
self.gui_progbar = gui_progbar
# Initialize Widgets
self.time_config_widget = GUI_TimeConfig(self.dirs, self.gui_progbar)
self.start_stop_btns_widget = GUI_StartStopBtns(self.dirs, self.gui_progbar)
self.save_config_widget = GUI_SaveConfig(self.dirs)
self.lj_graph_widget = GUI_LiveScrollingGraph(dirs)
self.lj_table_widget = GUI_LJDataReport()
self.lj_config_widget = GUI_LabJackConfig(self.dirs, self.lj_graph_widget)
self.ard_config_widget = GUI_ArdConfig(self.dirs, self.gui_progbar)
self.device_presets_widget = GUI_DevicePresets(self.dirs, self.gui_progbar,
self.lj_config_widget,
self.time_config_widget)
self.status_bars = GUI_StatusBars()
# Layout
self.setup_tabs()
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
self.add_to_grid()
def setup_tabs(self):
"""Sets up the main and config tabs"""
self.tabs = qg.QTabWidget(self)
tab_main = qg.QWidget(self.tabs)
tab_config = qg.QWidget(self.tabs)
self.tab_grid_main = qg.QGridLayout()
self.tab_grid_config = qg.QGridLayout()
tab_main.setLayout(self.tab_grid_main)
tab_config.setLayout(self.tab_grid_config)
self.tabs.addTab(tab_main, 'Main')
self.tabs.addTab(tab_config, 'Config')
self.add_to_tab_main()
self.add_to_tab_config()
def add_to_tab_main(self):
"""Add Widgets to Main Tab"""
self.tab_grid_main.addWidget(self.lj_graph_widget, 0, 0, 1, 5)
self.tab_grid_main.addWidget(self.status_bars, 1, 0, 1, 3)
self.tab_grid_main.addWidget(self.lj_table_widget, 1, 3, 1, 2)
def add_to_tab_config(self):
"""Add widgets to Config Tab"""
self.tab_grid_config.addWidget(self.ard_config_widget, 0, 0, 1, 2)
self.tab_grid_config.addWidget(self.lj_config_widget, 1, 1)
self.tab_grid_config.addWidget(self.device_presets_widget, 1, 0)
def add_to_grid(self):
"""Add tabs and non-tabbed widgets to grid"""
self.grid.addWidget(self.start_stop_btns_widget, 0, 0)
self.grid.addWidget(self.time_config_widget, 1, 0)
self.grid.addWidget(self.save_config_widget, 2, 0)
self.grid.addWidget(self.tabs, 0, 1, 3, 1)
def enable_disable_widgets(self, exp_running):
"""Enable or Disable Widgets depending on experiment state"""
# Widgets we enable on running an exp.
enabled_on_run = [self.start_stop_btns_widget.stop_btn]
# Widgets we disable on running an exp.
disable_on_run = [self.start_stop_btns_widget.start_btn,
self.start_stop_btns_widget.name_frame,
self.save_config_widget, self.time_config_widget,
self.lj_config_widget, self.ard_config_widget,
self.device_presets_widget]
# Config Enabled/Disabled:
for widget in enabled_on_run:
if exp_running:
widget.setEnabled(True)
else:
widget.setEnabled(False)
for widget in disable_on_run:
if exp_running:
widget.setEnabled(False)
else:
widget.setEnabled(True)
# Other changes
if exp_running:
self.lj_graph_widget.frame.setTitle('LabJack Live Stream (Low Frequency Scanning) - [Recording to File]')
if not exp_running:
self.lj_graph_widget.frame.setTitle('LabJack Live Stream (Low Frequency Scanning) - [Not Recording to File]')
class GUI_LabJackConfig(qg.QWidget):
"""Configuring LabJack Options"""
def __init__(self, dirs, grapher):
qg.QWidget.__init__(self)
self.num_ch = 14
self.dirs = dirs
self.lj_proc_updated = False
self.grapher = grapher
self.proc_handler_queue = PROC_HANDLER_QUEUE
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
# Setup and Layout
self.initialize()
def initialize(self):
"""Sets up Config Label and Controls"""
self.frame = qg.QGroupBox('LabJack Config')
self.grid.addWidget(self.frame)
grid = qg.QGridLayout()
self.frame.setLayout(grid)
# Create objects and Add to grid
grid.addWidget(self.init_summary_label(), 0, 0)
grid.addWidget(self.init_entry(), 1, 0)
grid.addWidget(self.init_checkboxes(), 2, 0)
self.reload_gui_info(True)
def init_summary_label(self):
"""Sets up a label for summarizing labjack settings"""
self.summ_label = qg.QLabel('')
self.summ_label.setAlignment(qc.Qt.AlignCenter)
self.summ_label.setFrameStyle(qg.QFrame.Sunken | qg.QFrame.StyledPanel)
return self.summ_label
def init_entry(self):
"""Sets up entry and button for scan freq"""
frame = qg.QFrame()
grid = qg.QGridLayout()
frame.setLayout(grid)
# Entry
freq = self.dirs.settings.lj_last_used.scan_freq
self.scan_freq_entry = GUI_IntOnlyEntry(max_digits=5, default_txt=str(freq))
self.scan_freq_entry.setText(str(freq))
# Max Freq Label
self.max_freq_label = qg.QLabel('')
self.max_freq_label.setAlignment(qc.Qt.AlignCenter)
self.max_freq_label.setFrameStyle(qg.QFrame.Sunken | qg.QFrame.StyledPanel)
self.set_max_freq_label()
# Buttons
confirm_btn = qg.QPushButton('Confirm')
confirm_btn.clicked.connect(self.save_scan_freq)
# Layout
grid.addWidget(qg.QLabel('Scan Frequency:'), 0, 0, 1, 4)
grid.addWidget(self.max_freq_label, 1, 0, 1, 5)
grid.addWidget(self.scan_freq_entry, 2, 0, 1, 4)
grid.addWidget(confirm_btn, 2, 4)
return frame
def init_checkboxes(self):
"""Sets up checkboxes for each channel"""
frame = qg.QFrame()
grid = qg.QGridLayout()
frame.setLayout(grid)
# Generate and Layout checkboxes
grid.addWidget(qg.QLabel('Channels Currently in Use:\n'), 0, 0, 1, 5)
self.checkboxes = [qg.QCheckBox('{:0>2}'.format(i)) for i in range(self.num_ch)]
[grid.addWidget(self.checkboxes[i],
i // 5 + 1,
i - (i // 5) * 5) for i in range(self.num_ch)]
[self.checkboxes[i].clicked.connect(self.save_channels) for i in list(range(self.num_ch))]
return frame
def set_max_freq_label(self):
"""Shows user the maximum scan freq allowed"""
num_ch = len(self.dirs.settings.lj_last_used.ch_num)
msg = 'Max Freq = 50000 / [{}] Channels = [{}] Hz' \
''.format(num_ch, int(50000 / num_ch))
self.max_freq_label.setText(msg)
def set_summ_label(self):
"""Sets the summary label to reflect most updated LJ settings"""
ch = self.dirs.settings.lj_last_used.ch_num
freq = self.dirs.settings.lj_last_used.scan_freq
self.summ_label.setText('Channels:\n{}\n\nScan Freq: [{} Hz]'.format(ch, freq))
def save_scan_freq(self):
"""Sets the scan frequency"""
scan_freq = self.scan_freq_entry.text().strip()
max_freq = int(50000 / len(self.dirs.settings.lj_last_used.ch_num))
if scan_freq == '' \
or int(scan_freq) > max_freq \
or int(scan_freq) == 0:
self.scan_freq_entry.visual_warning()
return
self.update_lj_last_used(scan_freq=int(deepcopy(scan_freq)), reset_gui_elements=False)
def update_lj_last_used(self, ch_num=None, scan_freq=None, send_to_proc_handler=True, reset_gui_elements=False):
"""Update dirs.settings.lj_last_used. Also notify proc_handler to update lj_proc settings"""
self.grapher.plots_are_reset = False
if not ch_num:
ch_num = deepcopy(self.dirs.settings.lj_last_used.ch_num)
if not scan_freq:
scan_freq = deepcopy(self.dirs.settings.lj_last_used.scan_freq)
# first notify proc_handler to update lj_proc settings
if not self.lj_proc_updated:
# Send Message
if send_to_proc_handler:
self.proc_handler_queue.put_nowait('{}{}|{}'.format(LJ_CONFIG, ch_num, scan_freq))
# check back every 5 ms until lj_proc has been updated
qc.QTimer.singleShot(5, lambda: self.update_lj_last_used(ch_num, scan_freq, False, reset_gui_elements))
# once proc_handler has updated lj_procs, we update the GUI
elif self.lj_proc_updated:
self.dirs.settings.lj_last_used.ch_num = deepcopy(ch_num)
self.dirs.settings.lj_last_used.scan_freq = deepcopy(scan_freq)
self.reload_gui_info(reset_gui_elements)
self.lj_proc_updated = False
self.grapher.update_graphs()
def save_channels(self):
"""Saves channels selected based on boxes checked"""
selected = [i for i in range(self.num_ch) if self.checkboxes[i].isChecked()]
max_freq = int(50000 / len(selected))
freq = self.dirs.settings.lj_last_used.scan_freq
if freq > max_freq:
freq = max_freq
self.scan_freq_entry.setText(str(freq))
self.update_lj_last_used(ch_num=selected, scan_freq=freq, reset_gui_elements=False)
def enable_disable_chkboxes(self):
"""Enable or disable check boxes depending on number of boxes checked"""
selected = [i for i in range(self.num_ch) if self.checkboxes[i].isChecked()]
# All check boxes available if selected between 1-7 channels
if 1 < len(selected) < 8:
[self.checkboxes[i].setEnabled(True) for i in range(self.num_ch)]
# Cannot select more than 8 channels
elif len(selected) == 8:
[self.checkboxes[i].setEnabled(False) for i in range(self.num_ch)
if i not in self.dirs.settings.lj_last_used.ch_num]
# Cannot select fewer than 1 channel
elif len(selected) == 1:
[self.checkboxes[i].setEnabled(False) for i in range(self.num_ch) if self.checkboxes[i].isChecked()]
def reload_gui_info(self, reset_gui_elements):
"""Changes label, entry, boxes checked based on settings file"""
if reset_gui_elements:
self.scan_freq_entry.setText(str(self.dirs.settings.lj_last_used.scan_freq))
self.set_channels()
self.set_summ_label()
self.enable_disable_chkboxes()
self.set_max_freq_label()
self.grapher.reset_plots()
def set_channels(self):
"""Sets channels based on which ones enabled in dirs.settings"""
[self.checkboxes[i].setChecked(True) for i in self.dirs.settings.lj_last_used.ch_num]
[self.checkboxes[i].setChecked(False) for i in range(self.num_ch)
if i not in self.dirs.settings.lj_last_used.ch_num]
class GUI_ArdConfig(qg.QWidget):
"""Configuring Arduino Input/Output Settings"""
def __init__(self, dirs, gui_progbar):
qg.QWidget.__init__(self)
self.dirs = dirs
self.gui_progbar = gui_progbar
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
# Setup and Layout
self.initialize()
self.setMaximumHeight(120)
self.grid.addWidget(self.frame)
def initialize(self):
"""Sets up buttons for arduino configuring"""
self.frame = qg.QGroupBox('Arduino Config')
grid = qg.QGridLayout()
self.frame.setLayout(grid)
# Dynamic "Types" Label
self.types_label = qg.QLabel(' ')
self.types_label.setAlignment(qc.Qt.AlignCenter)
self.types_label.setFrameStyle(qg.QFrame.Sunken | qg.QFrame.Panel)
self.types_label.setMaximumWidth(qt_text_metrics.width('Output'))
self.types_label.setMinimumWidth(qt_text_metrics.width('Output'))
# Pin Dropdown Menu
pins = ['', tone_pin, ''] + output_pins + [''] + pwm_pins
self.pins_dropdown = qg.QComboBox()
self.pins_dropdown.activated[str].connect(self.pins_dropdown_selection)
for pin in pins:
self.pins_dropdown.addItem(str(pin))
# Other Entries and Labels
static_labels = ['Pin', 'Type', 'Start Time', 'End Time', 'Freq', 'Phase Shift', 'Duty Cycle']
self.entries = {}
for index, label in enumerate(static_labels):
# Create Entries
if label not in ['Type', 'Pin']:
self.entries[label] = GUI_IntOnlyEntry(max_digits=6)
self.entries[label].setEnabled(False) # Initialize to Disabled
# Create Column Labels
static_labels[index] = qg.QLabel(label)
static_labels[index].setAlignment(qc.Qt.AlignCenter)
# Some special restrictions on phase shift and duty cycles
self.entries['Phase Shift'].setMaxLength(3)
self.entries['Phase Shift'].set_min_max_value(0, 360)
self.entries['Duty Cycle'].setMaxLength(2)
self.entries['Duty Cycle'].set_min_max_value(1, 99)
# Add to Grid
for index, label in enumerate(static_labels):
grid.addWidget(label, 0, index)
if label.text() not in ['Type', 'Pin']:
grid.addWidget(self.entries[label.text()], 1, index)
elif label.text() == 'Type':
grid.addWidget(self.types_label, 1, index)
elif label.text() == 'Pin':
grid.addWidget(self.pins_dropdown, 1, index)
# Confirm Button
self.confirm_btn = qg.QPushButton('\nConfirm\n')
self.confirm_btn.clicked.connect(self.add_new_config)
self.confirm_btn.setEnabled(False) # Initialize to Disabled
grid.addWidget(self.confirm_btn, 0, len(static_labels), 2, 1)
def pins_dropdown_selection(self, selection):
"""Changes widget output and enabled entries depending on pin selection"""
try:
selection = int(selection)
except ValueError:
self.types_label.setText('')
self.enable_disable_entries(types=None)
else:
# First we check if our pin selection is a selected arduino progbar
progbar_selected = self.gui_progbar.scene.selectedItems()
if not (len(progbar_selected) == 1 and selection == int(progbar_selected[0].data.pin)):
self.gui_progbar.reset_selection()
# Then we set the labels and enable or disable necessary entries
self.pins_dropdown.setCurrentIndex(self.pins_dropdown.findText(str(selection)))
if selection == tone_pin:
self.types_label.setText(tone)
self.enable_disable_entries(types=tone)
elif selection in output_pins:
self.types_label.setText(output)
self.enable_disable_entries(types=output)
elif selection in pwm_pins:
self.types_label.setText(pwm)
self.enable_disable_entries(types=pwm)
def enable_disable_entries(self, types):
"""Enables or Disables Entries based on field selection"""
universal_enabled = ['Start Time', 'End Time']
self.confirm_btn.setEnabled(True)
if not types:
enabled = []
self.confirm_btn.setEnabled(False)
elif types == tone:
enabled = universal_enabled + ['Freq']
elif types == output:
enabled = universal_enabled
elif types == pwm:
enabled = universal_enabled + ['Freq', 'Phase Shift', 'Duty Cycle']
for entry in self.entries:
if entry not in enabled:
self.entries[entry].setEnabled(False)
else:
self.entries[entry].setEnabled(True)
def add_new_config(self):
"""Adds a new visual and backend config for arduino settings based on user input"""
if not self.check_entries_valid():
return
new_config = ArdDataContainer(*self.get_entry_input())
# Adds the new config and reloads the progbar
# First we check if any conflicts exist
conflicts = self.check_new_config_conflicts(new_config)
# If we have selected exactly one progbar, then we make adjustments to it
# BUT: We make adjustments if and only if there are no conflicts or the conflict is with the selected bar
if len(self.gui_progbar.scene.selectedItems()) == 1 \
and ((len(conflicts) == 1 and conflicts[0] == self.gui_progbar.scene.selectedItems()[0].data)
or len(conflicts) == 0):
# We change the old config and overwrite it to the new config
selected_item_data = self.gui_progbar.scene.selectedItems()[0].data
index = self.dirs.settings.ard_last_used.configs.index(selected_item_data)
self.dirs.settings.ard_last_used.configs[index] = new_config
# Set the background
self.gui_progbar.set_dynamic_background()
# We then reselect the item so the user knows which bar they adjusted
for bar in self.gui_progbar.ard_stim_bars():
if bar.data == new_config:
bar.setSelected(True)
self.gui_progbar.setFocus()
# If no conflicts, we add the new config
elif len(conflicts) == 0:
self.dirs.settings.ard_last_used.configs.append(new_config)
self.gui_progbar.set_dynamic_background()
# Otherwise we send a warning to the user
else:
[item.visual_warning() for item in self.gui_progbar.ard_stim_bars()
for config in conflicts if item.data == config]
def check_entries_valid(self):
"""Checks that user inputs are appropriate inputs"""
# Are entries empty?
if not self.check_entries_not_empty():
return False
# Is the segment endtime greater than the segment start time?
elif int(self.entries['End Time'].text()) <= int(self.entries['Start Time'].text()):
self.entries['End Time'].visual_warning()
self.entries['Start Time'].visual_warning()
return False
# Are tone frequencies at least 50Hz?
elif self.types_label.text() == tone and int(self.entries['Freq'].text()) < 50:
self.entries['Freq'].visual_warning()
GUI_Message('Tone Frequencies must be at least 50Hz;\n\nUse PWM pins for Low Frequencies')
return False
# Are PWM frequencies at most 100Hz?
elif self.types_label.text() == pwm and int(self.entries['Freq'].text()) > 100:
self.entries['Freq'].visual_warning()
GUI_Message('PWM Frequencies must be at most 100Hz;\n\nUse Pin 10 (Tone) for High Frequencies')
return False
else:
return True
def check_new_config_conflicts(self, new_config):
"""Checks if a new user config conflicts with previous entries"""
conflicts = []
configs = self.dirs.settings.ard_last_used.configs
# Is new_config.pin already being used in a pre-existing config?
configs = [config for config in configs if new_config.pin == config.pin]
# If new_config.pin is already being used, does the new_config intersect with any previous configs?
if len(configs) != 0:
new_start, new_stop = new_config.time_on_ms, new_config.time_off_ms
time_segments = [(config.time_on_ms, config.time_off_ms) for config in configs]
conflicts = [configs[time_segments.index(segment)] for segment in time_segments
# New timepoints should not be within previous segments
if segment[0] < new_start < segment[1]
or segment[0] < new_stop < segment[1]
# New segments should not be within or encompass previous segments
or (segment[0] <= new_start and segment[1] >= new_stop)
or (segment[0] >= new_start and segment[1] <= new_stop)]
return conflicts
def check_entries_not_empty(self):
"""Checks that entries are not empty before pulling input from them"""
enabled_but_empty = [self.entries[entry]
for entry in self.entries
if self.entries[entry].isEnabled()
and self.entries[entry].text().strip() == '']
if len(enabled_but_empty) > 0:
[entry.visual_warning() for entry in enabled_but_empty]
return False
elif len(enabled_but_empty) == 0:
return True
def get_entry_input(self):
"""Gets user input from fields and returns specific parameters depending on type specified"""
# First we get the output of all fields
types = self.types_label.text()
pin = int(self.pins_dropdown.currentText())
time_on_ms = int(self.entries['Start Time'].text().strip()) * 1000
time_off_ms = int(self.entries['End Time'].text().strip()) * 1000
freq = self.entries['Freq'].text().strip()
phase_shift = self.entries['Phase Shift'].text().strip()
duty_cycle = self.entries['Duty Cycle'].text().strip()
# Then we distribute necessary information depending on types requested
if types == tone:
return time_on_ms, time_off_ms, types, 10, freq
elif types == output:
return time_on_ms, time_off_ms, types, pin
elif types == pwm:
return time_on_ms, time_off_ms, types, pin, freq, phase_shift, duty_cycle
def load_from_ard_bar(self, data):
"""Gets data from an implemented arduino bar in gui_progbar"""
# Parse data
pin, types, freq, phase_shift, duty_cycle, time_on, time_off = ('',) * 7
if data:
types = data.types
time_on = str(int(data.time_on_ms / 1000))
time_off = str(int(data.time_off_ms / 1000))
if data.types == tone:
pin = '10'
freq = str(data.freq)
elif data.types in [output, pwm]:
pin = str(data.pin)
if data.types == pwm:
freq = str(data.freq)
phase_shift = str(data.phase_shift)
duty_cycle = str(data.duty_cycle)
# Set Widgets
self.pins_dropdown.setCurrentIndex(self.pins_dropdown.findText(pin))
self.types_label.setText(types)
self.entries['Start Time'].setText(time_on)
self.entries['End Time'].setText(time_off)
self.entries['Freq'].setText(freq)
self.entries['Phase Shift'].setText(phase_shift)
self.entries['Duty Cycle'].setText(duty_cycle)
# Enable/Disable widgets
self.enable_disable_entries(types)
class GUI_DevicePresets(qg.QWidget):
"""Selecting/Saving User Defined Presets"""
def __init__(self, dirs, gui_progbar, lj_widget, time_widget):
qg.QWidget.__init__(self)
self.dirs = dirs
self.gui_progbar = gui_progbar
self.lj_widget = lj_widget
self.time_widget = time_widget
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
self.device_types = [arduino, labjack]
self.initialize()
def preset_names(self, device):
"""Returns list of presets depending on device;
we use a function since the list is dynamic, makes life easier"""
if device == labjack:
return [''] + [name for name in self.dirs.settings.lj_presets]
elif device == arduino:
return [''] + [name for name in self.dirs.settings.ard_presets]
def initialize(self):
"""Set up GUI elements"""
frame = qg.QGroupBox('Device Presets')
self.grid.addWidget(frame)
grid = qg.QGridLayout()
frame.setLayout(grid)
# QWidget Containers
self.entries = {dev: GUI_EntryWithWarning() for dev in self.device_types}
self.btns = {dev: qg.QPushButton('Save New') for dev in self.device_types}
self.dropdowns = {dev: qg.QComboBox() for dev in self.device_types}
# Set frame layout with two inner frames for each device's preset GUI
[grid.addWidget(self.init_preset_type(dev), i, 0) for i, dev in enumerate(self.device_types)]
self.connect_widgets()
def init_preset_type(self, dev):
"""Sets up an inner frame containing preset options depending on type passed"""
frame = qg.QFrame()
frame.setFrameStyle(qg.QFrame.Sunken | qg.QFrame.StyledPanel)
grid = qg.QGridLayout()
frame.setLayout(grid)
# Setup Widgets
grid.addWidget(qg.QLabel('{} Presets'.format(dev.capitalize())), 0, 0, 1, 5)
grid.addWidget(self.entries[dev], 2, 0, 1, 5)
grid.addWidget(self.btns[dev], 3, 2, 1, 1)
grid.addWidget(qg.QLabel('Select: '), 4, 0, 1, 1)
grid.addWidget(self.dropdowns[dev], 4, 1, 1, 4)
return frame
def connect_widgets(self):
"""Connects individual widgets to appropriate slots"""
for dev in self.device_types:
[self.dropdowns[dev].addItem(name) for name in sorted(self.preset_names(dev), key=str.lower)]
self.dropdowns[dev].activated[str].connect(lambda option, device=dev: self.select_preset(option, device))
self.btns[dev].clicked.connect(lambda cl, device=dev: self.save_preset(device))
def select_preset(self, option, device):
"""Based on user selection in dropdowns, implement the preset"""
if option == '':
return
if device == arduino:
self.dirs.settings.ard_last_used = deepcopy(self.dirs.settings.ard_presets[option])
self.gui_progbar.set_dynamic_background()
self.time_widget.set_text_in_entries()
elif device == labjack:
ch_num = self.dirs.settings.lj_presets[option].ch_num
scan_freq = self.dirs.settings.lj_presets[option].scan_freq
self.lj_widget.update_lj_last_used(ch_num=ch_num, scan_freq=scan_freq, reset_gui_elements=True)
def save_preset(self, device):
"""Saves user settings to a hardcopy preset"""
name = self.entries[device].text().strip()
# Is the entry empty?
if len(name) == 0:
self.entries[device].visual_warning()
return
# Is the entry overwriting a previous setting?
overwrite = None
if name in self.preset_names(device):
msg = '[{}]\nAlready exists as a preset!\nOverwrite anyway?'.format(name)
overwrite = qg.QMessageBox.question(self, 'Overwrite?', msg,
qg.QMessageBox.No | qg.QMessageBox.Yes,
qg.QMessageBox.No)
if overwrite == qg.QMessageBox.No:
return
# If not empty, and [choose to overwrite OR a new name], we proceed to save it.
if device == arduino:
self.dirs.settings.ard_presets[name] = deepcopy(self.dirs.settings.ard_last_used)
elif device == labjack:
self.dirs.settings.lj_presets[name] = deepcopy(self.dirs.settings.lj_last_used)
# Insert new preset name in the correct location;
# if we did an overwrite we don't need to add a new name
if overwrite != qg.QMessageBox.Yes:
ind = sorted(self.preset_names(device) + [name], key=str.lower).index(name)
self.dropdowns[device].insertItem(ind, name)
# Now we set the dropdown option to display our newly saved (or overwritten) option
self.dropdowns[device].setCurrentIndex(self.dropdowns[device].findText(name))
class GUI_StartStopBtns(qg.QWidget):
"""Start and Stop buttons and associated signals/slots.
Also contains an entry field for naming experiment runs"""
def __init__(self, dirs, gui_progbar):
qg.QWidget.__init__(self)
self.dirs = dirs
# Synchronization
self.proc_handler_queue = PROC_HANDLER_QUEUE
# GUI Objects
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
self.gui_progbar = gui_progbar
# Setup
self.setMinimumWidth(200)
self.setMaximumWidth(200)
self.init_btns()
self.init_name_entry()
self.add_to_grid()
def init_btns(self):
"""Creates and Connects Start Stop Buttons"""
self.start_btn = qg.QPushButton('START')
self.start_btn.setStyleSheet('background-color: cyan')
self.stop_btn = qg.QPushButton('STOP')
self.stop_btn.setStyleSheet('background-color: orange')
self.start_btn.clicked.connect(self.start_exp)
self.stop_btn.clicked.connect(self.stop_exp)
self.stop_btn.setEnabled(False)
def init_name_entry(self):
"""An entry field for user naming"""
self.name_frame = qg.QGroupBox('Trial Name:')
grid = qg.QGridLayout()
self.name_frame.setLayout(grid)
self.name_entry = GUI_EntryWithWarning()
grid.addWidget(self.name_entry)
def get_exp_name(self):
"""Gets the experiment name from user input; if empty, set to no_name"""
name = str(self.name_entry.text().strip())
if name == '':
return False
for i in name:
if i in FORBIDDEN_CHARS:
name = name.replace(i, '_')
return name
def add_to_grid(self):
"""Add Buttons to Grid"""
self.grid.addWidget(self.name_frame, 0, 0)
self.grid.addWidget(self.start_btn, 1, 0)
self.grid.addWidget(self.stop_btn, 2, 0)
def start_exp(self):
"""Starts the experiment"""
# Get the Exp name
name = self.get_exp_name()
if not name or name in self.dirs.list_file_names():
self.name_entry.visual_warning()
return
# Save directory checks and handling
self.dirs.check_dirs()
if not self.dirs.created_date_stamped_dir:
self.dirs.create_date_stamped_dir()
self.proc_handler_queue.put_nowait('{}{}'.format(DIR_TO_USE_HEADER, self.dirs.date_stamped_dir))
# Send start message to proc_handler
self.proc_handler_queue.put_nowait('{}{}'.format(RUN_EXP_HEADER, name))
def stop_exp(self):
"""Stops the experiment"""
self.proc_handler_queue.put_nowait(HARDSTOP_HEADER)
self.gui_progbar.stop_bar()
class GUI_TimeConfig(qg.QWidget):
"""Entries and Buttons for Configuring Experiment Time"""
def __init__(self, dirs, gui_progbar):
qg.QWidget.__init__(self)
self.proc_handler_queue = PROC_HANDLER_QUEUE
self.dirs = dirs
self.gui_progbar = gui_progbar
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
self.init_entries()
self.add_to_grid()
def init_entries(self):
"""Creates and Connects time config entries"""
self.entries = []
entry_types = [hh, mm, ss]
# Entries and Buttons (Signal Emitters)
hhmmss = time_convert(ms=self.dirs.settings.ttl_time())
for index, entry_type in enumerate(entry_types):
self.entries.append(GUI_IntOnlyEntry(max_digits=2, default_txt=str(hhmmss[index])))
self.entries[index].setMaximumWidth(qt_text_metrics.width('00000'))
self.set_text_in_entries()
self.ttl_time_confirm_btn = qg.QPushButton('\nOk\n')
self.ttl_time_confirm_btn.clicked.connect(self.confirm_time)
# Static Labels
static_labels = ['Hour', 'Min', 'Sec', ':', ':']
for index, name in enumerate(static_labels):
static_labels[index] = qg.QLabel(name)
static_labels[index].setAlignment(qc.Qt.AlignCenter)
# Keep time entries contained in its own frame
self.time_entry_frame = qg.QGroupBox()
self.time_entry_frame.setTitle('Total Experiment Time:')
self.time_entry_frame.setMaximumWidth(200)
grid = qg.QGridLayout()
self.time_entry_frame.setLayout(grid)
# Add components to frame
grid.addWidget(self.ttl_time_confirm_btn, 2, 0, 1, 5)
for index, entry in enumerate(self.entries):
grid.addWidget(static_labels[index], 0, index * 2)
grid.addWidget(entry, 1, index * 2)
if index < 2:
grid.addWidget(static_labels[index + 3], 1, index * 2 + 1)
def set_text_in_entries(self):
"""Sets the text based on dirs.settings"""
hhmmss = time_convert(ms=self.dirs.settings.ttl_time())
[self.entries[index].setText(str(hhmmss[index])) for index, entry_type in enumerate([hh, mm, ss])]
def confirm_time(self):
"""Sets the total experiment time"""
hhmmss = ''
# Get User Entry
for entry in self.entries:
hhmmss += '{:0>2}'.format(str(entry.text()))
# Error Checking
if int(hhmmss) < 5:
ms = 5000
else:
ms = time_convert(hhmmss=hhmmss)
# Is the time in the entries lower than the times configured in the prog bars?
if any([(ms < config.time_off_ms) for config in self.dirs.settings.ard_last_used.configs]):
ms = max([ms] + [config.time_off_ms for config in self.dirs.settings.ard_last_used.configs])
hhmmss = time_convert(ms=ms)
msg = 'Total time cannot be less than [{}:{}:{}] ' \
'because one of the arduino output endpoints exceeds this value.\n\n' \
'Reconfigure arduino outputs to ' \
'reduce total time.'.format(hhmmss[0], hhmmss[1], hhmmss[2])
GUI_Message(msg)
# Setting empty entries
for index, time in enumerate(time_convert(ms=ms)):
self.entries[index].setText(str(time))
self.dirs.settings.set_ttl_time(ms)
self.proc_handler_queue.put_nowait('{}{}'.format(TTL_TIME_HEADER, ms))
self.gui_progbar.set_dynamic_background()
def add_to_grid(self):
"""Add Widgets to Grid"""
self.grid.addWidget(self.time_entry_frame, 1, 0, 1, 2)
class GUI_SaveConfig(qg.QWidget):
"""User configurable save file names"""
def __init__(self, dirs):
qg.QWidget.__init__(self)
self.dirs = dirs
self.proc_handler_queue = PROC_HANDLER_QUEUE
self.grid = qg.QGridLayout()
self.setLayout(self.grid)
# Setup
self.setMaximumWidth(200)
self.initialize()
def initialize(self):
"""Gets list of existing save directories"""
grid = qg.QGridLayout()
frame = qg.QGroupBox('Current Save Directory:')
change_dir_btn = qg.QPushButton('Change Directory')
self.dir_label = qg.QLabel('')
self.set_dirs_label(self.dirs.settings.last_used_save_dir)
self.dir_label.setFrameStyle(qg.QFrame.Sunken | qg.QFrame.Panel)
# Connect Signals
change_dir_btn.clicked.connect(self.change_dirs_dialog)
# Layout
frame.setLayout(grid)
grid.addWidget(self.dir_label, 0, 0)
grid.addWidget(change_dir_btn, 1, 0)
self.grid.addWidget(frame)
def change_dirs_dialog(self):
"""Opens a dialog for user to specify save directory"""
directory = str(qg.QFileDialog.getExistingDirectory(None, "Select Directory",
self.dirs.settings.last_used_save_dir))
# Don't change anything if we cancelled the dialog
if len(directory) == 0:
return
# Otherwise set dir to the one we selected in dialog
self.dirs.settings.last_used_save_dir = directory
self.set_dirs_label(directory)
self.dirs.created_date_stamped_dir = False
def set_dirs_label(self, dirs):
"""Sets self.dirs_label with word wrapping on '\\' markers"""
max_len = 23
lines = []
curr_line = ''
for i in [d for d in dirs.split('\\')]:
if not len(curr_line) == 0:
curr_line += '\\'
if len(curr_line + i) < max_len:
curr_line += i
else:
lines.append(curr_line)
curr_line = '' + i
lines.append(curr_line) # Append the final line that hasn't been added yet
label = '\n'.join([''.join(line) for line in lines])
self.dir_label.setText(label)
| [
"tiange.l"
] | tiange.l |
7617b75c7a76b5009e6dca9b3e54dd53393f66d3 | b76876afc024f33491c0e9838366eb8934ae1e65 | /blog/tests.py | 8aae2a6a15cdd63918dad96cb0ab16cb648e644d | [] | no_license | wiremine/djang-modelfun | 03fb2ffc68a01822c7e767f9e51913be938d8a8f | a0188b53b8355839b78306ab9a8d7058cd00bafd | refs/heads/master | 2020-05-16T03:37:14.994587 | 2011-01-20T03:30:21 | 2011-01-20T03:30:21 | 977,476 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 747 | py | from django.test import TestCase
from django.test.client import Client
from django.core import mail
# mail.outbox will contain emails during tests
from django.http import HttpResponse
from django.core.urlresolvers import clear_url_caches
from django.conf import settings
class ControllerTestCase(TestCase):
def setUp(self):
pass
def tearDown(self):
pass
def test_controller_urls(self):
response = self.client.get('/blog/')
self.assertEqual(response.status_code, 200)
self.assertEqual(response.content, "Hello World")
response = self.client.get('/blog2/')
self.assertEqual(response.status_code, 200)
self.assertEqual(response.content, "Hello World")
| [
"wiremine@gmail.com"
] | wiremine@gmail.com |
e16ebb8614032c72852153165835f46bd16e6965 | 787407f3afc659a75915352368374670c265134f | /3dcnn_gwak/network.py | fa19e211c452b485ec440f4b9c7fd24b26478e2a | [] | no_license | swilso421/shiny-barnacle | 39c189d380488718cb56084225be705c7dce52fb | 01285a82fefa32f6ec06cd299c8bb994d593ab10 | refs/heads/master | 2020-03-18T04:09:26.326083 | 2018-07-18T20:29:17 | 2018-07-18T20:29:17 | 134,272,733 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,445 | py | import tensorflow as tf
NUM_CLASSES = 18
def CNN_3D(input_node, stddev = 0.03, name = '3DCNN'):
filter1 = (3, 3, 3, 1, 32)
filter2 = (3, 3, 3, 32, 64)
filter3 = (3, 3, 3, 64, 128)
filter4 = (3, 3, 3, 128, 256)
weights1 = tf.Variable(tf.truncated_normal(filter1, stddev = stddev), name = name + '_W1')
weights2 = tf.Variable(tf.truncated_normal(filter2, stddev = stddev), name = name + '_W2')
weights3 = tf.Variable(tf.truncated_normal(filter3, stddev = stddev), name = name + '_W3')
weights4 = tf.Variable(tf.truncated_normal(filter4, stddev = stddev), name = name + '_W4')
weights5 = tf.Variable(tf.truncated_normal([2048, 1024], stddev = stddev), name = name + '_W5')
weights6 = tf.Variable(tf.truncated_normal([1024, 1024], stddev = stddev), name = name + '_W6')
weights7 = tf.Variable(tf.truncated_normal([1024, NUM_CLASSES], stddev = stddev), name = name + '_W7')
bias1 = tf.Variable(tf.truncated_normal(32), name = name + '_B1')
bias2 = tf.Variable(tf.truncated_normal(64), name = name + '_B2')
bias3 = tf.Variable(tf.truncated_normal(128), name = name + '_B3')
bias4 = tf.Variable(tf.truncated_normal(256), name = name + '_B4')
bias5 = tf.Variable(tf.truncated_normal(1024), name = name + '_B5')
bias6 = tf.Variable(tf.truncated_normal(1024), name = name + '_B6')
bias7 = tf.Variable(tf.truncated_normal(NUM_CLASSES), name = name + '_B7')
window = [1, 2, 2, 2, 1]
#Convolutional Layer #1
conv1 = tf.nn.conv3d(input_node,
filter = weights1,
strides = [1, 1, 1, 1, 1],
padding = 'SAME'
name = name + '_CONV1')
conv1 += bias1
relu1 = tf.nn.relu(conv1, name = name + '_RELU1')
pool1 = tf.nn.max_pool3d(relu1,
ksize = window,
strides = window,
padding = 'SAME',
name = name + '_POOL1')
#Convolutional Layer #2
conv2 = tf.nn.conv3d(pool1,
filter = weights2,
strides = [1, 1, 1, 1, 1],
padding = 'SAME'
name = name + '_CONV2')
conv2 += bias2
relu2 = tf.nn.relu(conv2, name = name + '_RELU2')
pool2 = tf.nn.max_pool3d(relu2,
ksize = window,
strides = window,
padding = 'SAME',
name = name + '_POOL2')
#Convolutional Layer #3
conv3 = tf.nn.conv3d(pool2,
filter = weights3,
strides = [1, 1, 1, 1, 1],
padding = 'SAME'
name = name + '_CONV3')
conv3 += bias3
relu3 = tf.nn.relu(conv3, name = name + '_RELU3')
pool3 = tf.nn.max_pool3d(relu3,
ksize = window,
strides = window,
padding = 'SAME',
name = name + '_POOL3')
#Convolutional Layer #4
conv4 = tf.nn.conv3d(pool3,
filter = weights4,
strides = [1, 1, 1, 1, 1],
padding = 'SAME'
name = name + '_CONV4')
conv4 += bias4
relu4 = tf.nn.relu(conv4, name = name + '_RELU4')
pool4 = tf.nn.max_pool3d(relu4,
ksize = window,
strides = window,
padding = 'SAME',
name = name + '_POOL4')
flattened = tf.reshape(pool4, [-1, 2048])
fc5 = tf.nn.matmul(flattened, weights5) + bias5
fc5_relu = tf.nn.relu(fc5, name = name + '_FC5')
if 0.0 < dropout < 1.0:
fc5_relu = tf.nn.dropout(fc5_relu, dropout, name = name + '_DROP1')
fc6 = tf.nn.matmul(fc5_relu, weights6) + bias6
fc6_relu = tf.nn.relu(fc6, name = name + '_FC6')
if 0.0 < dropout < 1.0:
fc6_relu = tf.nn.dropout(fc6_relu, dropout, name = name + '_DROP2')
fc7 = tf.nn.matmul(fc6_relu, weights7) + bias7
return fc7
| [
"swilso421@gmail.com"
] | swilso421@gmail.com |
0cb89a8adfed56ffbd5bc7113d6cc6a65c390491 | a0c130ab410d2ceebe8b8b28a80cb95a3f33f35c | /aei2/locweat.py | d0cf84e76d6ca9467740c4c393a9c69e6987fe09 | [] | no_license | zkevinfu/emotion-simulator | 1180dd7f5694ceae1c38ad291d49143950fdbce0 | 46679fbc18d7640c808eff388da56700cad07c24 | refs/heads/master | 2020-03-30T02:36:06.748842 | 2018-09-28T00:23:51 | 2018-09-28T00:23:51 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,580 | py | import requests
import Queue
import json
def get_loc():
loc_url = 'http://freegeoip.net/json'
r = requests.get(loc_url)
j = json.loads(r.text)
return (j['latitude'], j['longitude'])
def get_weat():
ds_url_base = 'https://api.darksky.net/forecast'
ds_api_key = '12caf1dbd09b0e555c8b4a27339abb75'
exclude = '?exclude=minutely,hourly,daily,flags,alerts'
lat_lon = get_loc()
latitude, longitude = lat_lon[0], lat_lon[1]
ds_request = '%s/%s/%s,%s%s'%(ds_url_base, ds_api_key,latitude,longitude,exclude)
ds_r = requests.get(ds_request)
ds_j = json.loads(ds_r.text)
return ds_j['currently']
def get_stim():
ds_j = get_weat()
weat = ds_j['icon']
temp = ds_j['temperature']
clear = ['clear-day', 'clear-night', 'wind']
p_cloudy = ['partly-cloudy-day', 'partly-cloudy-night']
cloudy = ['cloudy', 'fog']
rain = ['rain', 'sleet']
snow = ['snow']
e_queue = Queue.Queue()
#TODO finish this
if (any(weat == icon for icon in clear)):
e_queue.put(('joy', 1))
elif (any(weat == icon for icon in p_cloudy)):
e_queue.put(('sadness', 1))
elif (any(weat == icon for icon in cloudy)):
e_queue.put(('sadness', 1.5))
elif (any(weat == icon for icon in rain)):
e_queue.put(('sadness', 1.5))
elif (any(weat == icon for icon in snow)):
e_queue.put(('joy', 1))
if 40 <= temp <= 80:
e_queue.put(('joy', 1))
elif temp < 40:
e_queue.put(('sadness', 1.5))
elif temp > 80:
e_queue.put(('anger', 1.5))
return e_queue
| [
"zkevinfu@gmail.com"
] | zkevinfu@gmail.com |
486782999f26b2d806dfe00600731c63363f0469 | 5f9d274602d630173ac36227d8a70ed163d7a8d6 | /quickroute/utils.py | aeda2a4660fbeb2a7cdd6960a6d3538fc3826cc8 | [
"Apache-2.0"
] | permissive | jvtm/python-quickroute | 88ef1f594f6da99e1300bf36e6bdc1b16d6ab752 | 758e97cf38200efd649cb797c64c51b608db5de5 | refs/heads/master | 2020-03-30T23:06:24.348614 | 2014-08-14T20:37:54 | 2014-08-14T20:37:54 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 465 | py | '''
Python QuickRoute utility functions
'''
from datetime import datetime, timedelta
import json
class DateTimeEncoder(json.JSONEncoder):
""" Replacement JSON encoder supporting datetime and timedelta objects """
def default(self, obj):
if isinstance(obj, datetime):
return obj.isoformat()
elif isinstance(obj, timedelta):
return str(obj)
else:
return super(DateTimeEncoder, self).default(obj)
| [
"jvtm@kruu.org"
] | jvtm@kruu.org |
343ee30c9eeda3d263414b6da3a2aaaee932a506 | 74bd3a29425a6dae085af292d71b0bcbee5a0744 | /music_eda_STFT.py | b1246e477fffb8439fddcd019a0936dc6ce1720e | [] | no_license | tacyi/graduation_p | 3d42746d3e87f08754203a0a9fec19df95297324 | ce9112537165f8cd40f4253d938ffef8cd1b4728 | refs/heads/master | 2020-11-27T17:10:57.098501 | 2019-12-22T08:08:13 | 2019-12-22T08:08:13 | 229,536,101 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,672 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
# @Time : 2019/12/3 18:38
# @Author : tacyi
import matplotlib.pyplot as plt
import numpy as np
import scipy.signal as signal
from scipy.io import wavfile
import os
# 画短时傅里叶声谱图并保存图片,代码相关参考链接https://blog.csdn.net/qq7835144/article/details/88887576,
# 最主要的疑问是我不确定
# (sample_rate, x) = wavfile.read(rd)中的总采样数x是否是
# signal.stft(x, **params)的参数x;官方给了例子我看的很懵懂。
# 官方链接https://docs.scipy.org/doc/scipy/reference/generated/scipy.signal.stft.html#scipy.signal.stft
# 下面是我的实现
def stft_specgram(g, n, **params):
rd = 'C:/Users/tacyi/Desktop/genres/' + g + '/' + g + '.' + str(n).zfill(5) + '.wav'
# 主要深究的函数有下面几个
(sample_rate, x) = wavfile.read(rd)
# f: 采样频率数组;t: 时间数组;Zxx: STFT结果
f, t, zxx = signal.stft(x, **params)
plt.pcolormesh(t, f, np.abs(zxx))
plt.colorbar()
plt.title('STFT Magnitude')
plt.ylabel('Frequency [Hz]')
plt.xlabel('Time [sec]')
# plt.tight_layout(pad=0)
plt.tight_layout()
sd = './picture/STFT/' + g
if not os.path.exists(sd):
os.makedirs(sd)
sp = './picture/STFT/' + g + '/' + g + '.' + str(n).zfill(5) + '.png'
# if sp is not None:
if not os.path.exists(sp):
plt.savefig(sp) # 保存图像
plt.clf() # 清除画布
# return t, f, zxx
genre_list = ['blues', 'classical', 'country', 'disco', 'hiphop', 'jazz', 'metal', 'pop', 'reggae', 'rock']
for g in genre_list:
for n in range(100):
stft_specgram(g, n)
| [
"tacyi@139.com"
] | tacyi@139.com |
ae79c1df65cf49985b7a5950ad1c6f7c40f0cad6 | 3c80d771e2ca592a54db7b7a07e1d75180cf423d | /TestFolder/TestClasses.py | d033c998f4426f1cf07f305b681d683ff6e0310b | [] | no_license | TobiasTiedtke/Projekt_dsd_gui | 4873c759739f1011fd473eb69d1426225f123d21 | 83c02efd8776353db003f23404b25dc52df8bb33 | refs/heads/master | 2020-08-26T16:49:35.754859 | 2020-01-29T10:21:43 | 2020-01-29T10:21:43 | 217,078,904 | 0 | 0 | null | 2020-01-29T13:05:13 | 2019-10-23T14:32:50 | Python | UTF-8 | Python | false | false | 436 | py | import os, sys
def SingleBrowse():
#browsing for a folder and changing it to a string
filePath = "/informatik2/students/home/6kornell/Schreibtisch/bitbots_behavior-master/bitbots_head_behavior/src/bitbots_head_behavior/actions/look_at.py"
f = open(filePath, 'r')
for line in f:
line = str(line)
if line.startswith("class"):
line = line.split(" ")[1]
line = line.split("(")[0]
print(line)
SingleBrowse()
| [
"fwwx652@informatik.uni-hamburg.de"
] | fwwx652@informatik.uni-hamburg.de |
55eda8146943aa974f2b0b70a0e47c6a5dbd6c54 | da013048a61c4e34419c0b7ec721203ab1081173 | /large_shirts.py | f01231ceec2dadb643abac4273e53785d9a5964c | [] | no_license | makeTaller/Crash_Course_Excercises | 2c3bd3ac7c50a1830ce087b8187b93d93fab771e | 66b50906e6b33e6ba0d15a8a98ee17ca689392c4 | refs/heads/master | 2020-03-18T04:17:42.342365 | 2018-07-17T22:02:45 | 2018-07-17T22:02:45 | 134,280,608 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 220 | py | size = [ "small","medium","large"]
def large_shirts(message="I love python",size="large"):
print("Your shirt will be " + size + " and will read " + message)
large_shirts()
large_shirts("I Love the world!","medium")
| [
"kirktolliver28@gmail.com"
] | kirktolliver28@gmail.com |
3ffdef70c90547fe0a5a46765f97abf944b606a2 | 2ebbeef48fc7094f2089c5a82119d02af65a82dc | /codegolf/temperatures/solution.py | 1131f415d0ff594a97c6a114b11d5ce49eea0708 | [] | no_license | marty-Wallace/CodinGame | b93812c253795a3fba8ecc3bfa01f6f6b5047d26 | 6eeb5b5d1cff15c475cc2a7b6b8bcac68367b9fd | refs/heads/master | 2020-07-01T22:07:44.289785 | 2017-03-01T04:59:45 | 2017-03-01T04:59:45 | 74,251,457 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 75 | py | input()
print(min(map(int,input().split()),key=lambda x:x*x-x/9,default=0)) | [
"martin.v.wallace@ieee.org"
] | martin.v.wallace@ieee.org |
84c66116e8584f0c2d0ad42b6fdb367629748331 | a51453f9d5f7a93d5ce7f909e96507d24a34b40c | /small_knowledge.py/extract_num.py | e8159d0681311895eb5599a7cf0383cf6312cf4b | [] | no_license | justkissme/crawler | 99a228019836700c4cdb548d6ab76f3e3b1c0ddb | b0a6964d657486f6af90519f885352945e107409 | refs/heads/master | 2021-03-22T15:57:46.195433 | 2020-03-15T01:49:10 | 2020-03-15T01:49:10 | 247,380,355 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 152 | py | link = 'https://www.xslou.com/yuedu/9356/'
id_list = list(filter(str.isdigit,link))
book_id = ''.join(id_list)
print(type(book_id))
print(book_id) | [
"noreply@github.com"
] | noreply@github.com |
d4a5a2155aa71f6f81e1301fb6dea5d302b0742f | 9743d5fd24822f79c156ad112229e25adb9ed6f6 | /xai/brain/wordbase/nouns/_greens.py | ce3b23baa678edd94ee5bf830fa189133e5ffadb | [
"MIT"
] | permissive | cash2one/xai | de7adad1758f50dd6786bf0111e71a903f039b64 | e76f12c9f4dcf3ac1c7c08b0cc8844c0b0a104b6 | refs/heads/master | 2021-01-19T12:33:54.964379 | 2017-01-28T02:00:50 | 2017-01-28T02:00:50 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 231 | py |
from xai.brain.wordbase.nouns._green import _GREEN
#calss header
class _GREENS(_GREEN, ):
def __init__(self,):
_GREEN.__init__(self)
self.name = "GREENS"
self.specie = 'nouns'
self.basic = "green"
self.jsondata = {}
| [
"xingwang1991@gmail.com"
] | xingwang1991@gmail.com |
8bbd1f616dd60b38607f42fb53b0bed2badfe9b5 | 583f91afcb377707c830ad739b3d22d38f7eb4bf | /venv/bin/easy_install | d37d504868e9c4dc0754e20af384f4987ab2648d | [] | no_license | wjwalcher/CSP | 1544556606ea83f54aaceff9ad701c8eb4bedb7a | 1bddfd6dc5d70350323cd60224df804ef67dd581 | refs/heads/master | 2020-04-01T12:49:25.271546 | 2018-10-16T05:03:35 | 2018-10-16T05:03:35 | 153,225,369 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 448 | #!/Users/williamwalcher/PycharmProjects/322CSP/venv/bin/python
# EASY-INSTALL-ENTRY-SCRIPT: 'setuptools==39.1.0','console_scripts','easy_install'
__requires__ = 'setuptools==39.1.0'
import re
import sys
from pkg_resources import load_entry_point
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(
load_entry_point('setuptools==39.1.0', 'console_scripts', 'easy_install')()
)
| [
"williamwalcher@Williams-MacBook-Pro-3.local"
] | williamwalcher@Williams-MacBook-Pro-3.local | |
2e7cbcd4794cba57103a59e55b86743311629cf7 | 9ed58ef096d558a024aef9927e7f53fdf8805f85 | /applications/expensive_seq/expensive_seq.py | a599aede02737774b47914d10e7b5686d1e4d3e6 | [] | no_license | squashgray/Hash-Tables | 11e6c4e0cb9aa0dd3a2592fd04918311e5a23346 | d97ab7dad3dd34277b8084badcd27f19cb848ee6 | refs/heads/master | 2022-10-26T11:24:43.664825 | 2020-06-10T04:34:42 | 2020-06-10T04:34:42 | 261,280,162 | 1 | 0 | null | 2020-05-07T03:31:35 | 2020-05-04T19:46:26 | null | UTF-8 | Python | false | false | 223 | py | def expensive_seq(x, y, z):
# Implement me
if __name__ == "__main__":
for i in range(10):
x = expensive_seq(i*2, i*3, i*4)
print(f"{i*2} {i*3} {i*4} = {x}")
print(expensive_seq(150, 400, 800))
| [
"beej@beej.us"
] | beej@beej.us |
6be8285fe34928e223d9dcc5d0e598306f563657 | c762138fc50d2ac66a3fa9314102c9f15c8e0a2e | /S3utils.py | 1029978ca960245b77350ddc4dce7bffa9444663 | [] | no_license | saivamsikayala/garbage-detection-flask | 500492d013a47196dd04acc490ec955893a0560c | 5d380f4468af5de6290e19f555cf5b1223f6495f | refs/heads/master | 2022-11-30T10:10:22.134815 | 2020-08-01T09:37:32 | 2020-08-01T09:37:32 | 283,982,797 | 0 | 1 | null | 2020-08-01T04:30:49 | 2020-07-31T08:24:24 | Python | UTF-8 | Python | false | false | 1,180 | py | import boto3
import uuid
import io
from . import settings
class S3utils():
def __init__(self):
self._session = boto3.session.Session()
self._client = self._session.client('s3',
region_name = settings.AWS_S3_REGION_NAME,
endpoint_url = settings.AWS_S3_ENDPOINT_URL,
aws_access_key_id = settings.AWS_ACCESS_KEY_ID,
aws_secret_access_key = settings.AWS_SECRET_ACCESS_KEY)
def getfilename(self):
return f"smartbins/{str(uuid.uuid4())}.jpg"
def upload(self, filepath):
filename = self.getfilename()
#data = pilimage #self.toByteArray(pilimage)
#print(settings.AWS_STORAGE_BUCKET_NAME, filename)
#self._client.put_object(ACL='public_read', Bucket=settings.AWS_STORAGE_BUCKET_NAME, Key=filename, Body=data)
self._client.upload_file(filepath, settings.AWS_STORAGE_BUCKET_NAME, filename, ExtraArgs={'ACL': 'public-read'})
return f'{settings.AWS_S3_ENDPOINT_URL}/{settings.AWS_STORAGE_BUCKET_NAME}/{filename}'
def delete(self, filename):
return self._client.delete_object(Bucket='SIH2020', Key=filename)
S3Connection = S3utils() | [
"abhinav.dayal@hotmail.com"
] | abhinav.dayal@hotmail.com |
5c9d3d6c5f48bd56c8f41c7340f23a939dd6540c | 7bf59a25a66578cff650b0435ac95485ff51b43f | /model/couresDB.py | 4e1b59b80a6807a10bccdb4571b6ca3412fc9c41 | [] | no_license | amekaki/cmajorMiniProgram | a02ee5e9fcf93922e13c15ce9d2504211fc4324f | 15517e6d5c8514f8ac262265b77f2e60a0b3b9b9 | refs/heads/master | 2023-03-10T04:10:20.036776 | 2020-12-11T02:49:59 | 2020-12-11T02:49:59 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 178 | py | # -*- coding: UTF-8 -*-
from model.modelDB import Course
def getAllCourses():
res = Course.query.all()
return list(map(lambda x: {"cId": x.cId, "cname": x.cname}, res))
| [
"1150774341@qq.com"
] | 1150774341@qq.com |
6e2274d661b06ca37349193c0aebe92de6199296 | e7a79c157fe0edde0cabb4cebc707ce1530b5b7e | /answers/cartridge_base_state.py | e281bddc7ae481e400eea459831529801ab1ff55 | [] | no_license | Logsod/noti_rest_server | 1684d1d03c67aea46b4c6888d0951477de2888af | 529e9a0c66a6c7021224a2daf60f378f01164ca5 | refs/heads/master | 2023-08-10T16:39:43.000835 | 2021-09-29T13:19:41 | 2021-09-29T13:19:41 | 408,777,957 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,381 | py | import falcon
from flask import json
import db
class CartridgeBaseState:
def on_post(self, req, resp):
"""handle post"""
param = req.media
action = int(param['action'])
data = db.Db()
if action == 0: # add
if "cartridge_id" in param and "amount" in param:
result = data.addCartridgesToBaseState(param)
resp.status = falcon.HTTP_200
resp.content_type = falcon.MEDIA_TEXT
resp.text = result
elif action == 1: # change amount
data.changeBaseStateCartridgeAmount(param)
resp.status = falcon.HTTP_200
resp.content_type = falcon.MEDIA_TEXT
elif action == 2: # take one cartridge (insert into state table list and set status to 1) 1 equal work status
data.takeOneCartridge(param)
resp.status = falcon.HTTP_200
resp.content_type = falcon.MEDIA_TEXT
elif action == 3: # delete
data.deleteCartridgeFromBaseState(param['id'])
resp.status = falcon.HTTP_200
resp.content_type = falcon.MEDIA_TEXT
def on_get(self, req, resp):
data = db.Db()
result = data.getAllBaseStateCartridges()
print(result)
resp.status = falcon.HTTP_200
resp.content_type = falcon.MEDIA_TEXT
resp.text = result
| [
"wwwref@gmail.com"
] | wwwref@gmail.com |
34612b6bfe10f23da813bdc293d433a1103a70bd | ad9eaab5b382190ce76fb921c1a12f1c3adc4e09 | /main.py | 5993a711e30fa1537e727e5512bebe83a5847a92 | [] | no_license | dex1cre/asistent_bot | 9e7bc1ed3a293e1eea77fea258aafcafd469531e | 9b93166117a600815d45ab2b05940fdd9e065747 | refs/heads/master | 2021-01-21T10:04:34.188968 | 2017-03-07T19:30:36 | 2017-03-07T19:30:36 | 83,360,574 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 8,185 | py | #!/usr/bin/python3
#-*- coding: utf-8 -*-
#главный(здесь) модуль для работы с telegram
import telebot as tlb
#модуль для определения дня недели
from datetime import datetime
#модуль для многопоточности
import threading as potok
#модуль для работы с базами данных sqlite3
import sqlite3
#модуль с основными константами
import config
#constant
import string
#for pause
import time
bot = tlb.TeleBot(config.token)
def pr_mn(tk):
con = sqlite3.connect(config.url)
cur = con.cursor()
sql = "SELECT * FROM users WHERE id_user=" + str(tk)
try:
st = cur.execute(sql)
except sqlite3.DatabaseError as err:
print("Ошибка: " + err)
else:
try:
st = st.fetchone()
t = st[1]
except:
return "bad"
else:
return "good"
cur.close()
con.close()
def plans(text):
first = 0
second = 0
text = text[::-1]
#first symbol is number?
try:
first = int(text[0])
except:
return False
else:
#try to get second number
st = ""
for i in text:
try:
g = int(i)
except:
if i == "\n":
second = int(st[::-1])
st = ""
ind = text.index(i) + 1
for i in range(ind, len(text)):
try:
g = int(text[i])
except:
if st != "":
first = int(st[::-1])
return (first, second, -1 * i)
else:
return False
else:
st = st + text[i]
else:
return False
else:
st = st + i
def mn():
try:
@bot.message_handler(commands=["start"])
def send_to_start(message):
if pr_mn(message.from_user.id) == "good":
print(message.from_user.id, "перезапустил сессию")
else:
print(message.from_user.id, "создал первую сессию")
con = sqlite3.connect(config.url)
cur = con.cursor()
sql = "INSERT INTO users(id_user) VALUES('" + str(message.from_user.id) + "')"
try:
cur.execute(sql)
except sqlite3.DatabaseError as err:
print("Ошибка ", err)
else:
print(message.from_user.id, " добавлен в базу данных")
con.commit()
cur.close()
con.close()
um = tlb.types.ReplyKeyboardMarkup(True)
um.row("/start", "/now", "/stop")
um.row("/new", "/snew", "/week")
bot.send_message(message.from_user.id, "Начнём!", reply_markup=um)
#Завершение сессии
@bot.message_handler(commands=["stop"])
def send_to_stop(message):
config.wt = False
hm = tlb.types.ReplyKeyboardRemove()
bot.send_message(message.from_user.id, "До встречи!", reply_markup=hm)
#what's now?
@bot.message_handler(commands=["now"])
def send_to_stop(message):
print("The user with id: " + str(message.from_user.id) + " use the command NOW")
idd = message.from_user.id
now = datetime.now()
id_day = now.weekday()
print(id_day)
sql = "SELECT * FROM plans WHERE id_user=" + str(idd) + " AND id_day=" + str(id_day)
con = sqlite3.connect(config.url)
cur = con.cursor()
try:
t = con.execute(sql).fetchall()
except sqlite3.DatabaseError as err:
print(err)
bot.send_message(message.from_user.id, "some Error, ssory =)\nnow is " + str(id_day+1))
else:
st = ""
for i in t:
st = st + str(i[1]) + "time start: " + str(i[3]) + "\ntime stop: " + str(i[4]) + "\n---------------\n"
bot.send_message(message.from_user.id, st)
cur.close()
con.close()
#new
@bot.message_handler(commands=["new"])
def send_to_new(message):
config.wt = True
print("The user with id: " + str(message.from_user.id) + " use the command NEW")
st = "Чтобы добавить новые задачи напишите номер дня недели от 1 до 7"
bot.send_message(message.from_user.id, st)
#stop plans
@bot.message_handler(commands=["snew"])
def send_to_stop(message):
print("The user with id: " + str(message.from_user.id) + " use the command SNEW")
config.wt = False
bot.send_message(message.from_user.id, "Вы закончили писать задания,\n вы молодец!")
#text with commands
@bot.message_handler(content_types=["text"])
def send_to_text(message):
print(config.number)
if config.wt and message.text in config.numbers:
config.number = int(message.text)
bot.send_message(message.from_user.id, "Вы выбрали " + config.days[int(message.text)-1] + config.write_plans)
elif config.wt:
bl = plans(message.text)
tp = type(bl)
if tp == tuple:
idd = message.from_user.id
text = message.text
start = bl[0]
stop = bl[1]
ms = text[:bl[2]]
con = sqlite3.connect(config.url)
cur = con.cursor()
number = config.number-1
st = (ms, idd, start, stop, number)
sql = "INSERT INTO plans(what, id_user, dtime_start, dtime_stop, id_day) VALUES(?, ?, ?, ?, ?)"
try:
cur.execute(sql, st)
except sqlite3.DatabaseError as err:
print(err)
else:
con.commit()
print(ms)
bot.send_message(message.from_user.id, config.write_more)
elif tp == bool and not bl:
bot.send_message(message.from_user.id, "Что-то с вашим заданием не так! Проверьте, правильно ли вы вводите время начала и конца!")
else:
bot.send_message(message.from_user.id, "Nope, 1 2 3 4 5 6 7")
except:
print("Warning!")
#clear plans
def clear_plans():
now = datetime.now()
day = now.weekday()
while True:
if day == 6 and not config.week_none:
config.week_none = True
if day == 0 and config.week_none:
config.week_none = False
con = sqlite3.connect(config.url)
sql = "SELECT * FROM users"
t = cur.execute(t).fetchall()
for i in t:
bot.send_message(i[1], "Через минуту будет произведена отчистка данных прошлой недели, \nпожалуйста не вводите новых данных, иначе они будут утеряны")
time.sleep(60)
cur = con.cursor()
sql = "DELETE FROM plans"
cur.execute(sql)
con.commit()
cur.close()
con.close()
print("plans are cleared")
time.sleep(60)
t1 = potok.Thread(target=mn)
t2 = potok.Thread(target=clear_plans)
t1.daemon = True
t2.daemon = True
t1.start()
t2.start()
bot.polling(none_stop=True, interval=0)
| [
"dex1cre@mail.ru"
] | dex1cre@mail.ru |
780ef530a2c311710a82743db16d99d91edf5d3b | 35387a94b884466188c525a11cf7bf4dbf3c291f | /deepspeed/pt/deepspeed_constants.py | 1ba3dd482e25b1782b88a86ceab5d7a308f7bf4d | [
"MIT",
"LicenseRef-scancode-generic-cla"
] | permissive | liuyq47/DeepSpeed | 6808653e9adf490799fe9fe630d8cf20bc4a51b2 | cd68e6e55a2155ffaea681b18e012aafd686dce4 | refs/heads/master | 2022-12-12T07:35:00.362124 | 2020-08-10T17:09:40 | 2020-08-10T17:09:40 | 278,738,026 | 0 | 0 | MIT | 2020-08-26T16:14:54 | 2020-07-10T21:37:45 | Python | UTF-8 | Python | false | false | 8,682 | py | """
Copyright (c) Microsoft Corporation
Licensed under the MIT license.
"""
#############################################
# Routes
#############################################
ROUTE_TRAIN = "train"
ROUTE_EVAL = "eval"
ROUTE_PREDICT = "predict"
ROUTE_ENCODE = "encode"
#############################################
# Batch size
#############################################
TRAIN_BATCH_SIZE = "train_batch_size"
TRAIN_BATCH_SIZE_DEFAULT = None
#############################################
# Optimizer and lr scheduler
#############################################
OPTIMIZER = "optimizer"
OPTIMIZER_TYPE_DEFAULT = None
OPTIMIZER_PARAMS = "params"
TYPE = "type"
LEGACY_FUSION = "legacy_fusion"
LEGACY_FUSION_DEFAULT = False
SCHEDULER = "scheduler"
SCHEDULER_TYPE_DEFAULT = None
SCHEDULER_PARAMS = "params"
MAX_GRAD_NORM = 'max_grad_norm'
#############################################
# Optimizer and lr scheduler
#############################################
ZERO_ALLOW_UNTESTED_OPTIMIZER = "zero_allow_untested_optimizer"
ZERO_ALLOW_UNTESTED_OPTIMIZER_DEFAULT = False
#############################################
# Torch distributed constants
#############################################
TORCH_DISTRIBUTED_DEFAULT_PORT = "29500"
# Steps
STEPS_PER_PRINT = "steps_per_print"
STEPS_PER_PRINT_DEFAULT = 10
#########################################
# Training micro batch size per GPU
#########################################
# Batch size for one training step. This is used when the
# TRAIN_BATCH_SIZE cannot fit in GPU memory to determine
# the number of gradient accumulation steps. By default, this
# is set to None. Users can configure in ds_config.json as below example:
TRAIN_MICRO_BATCH_SIZE_PER_GPU = '''
TRAIN_MICRO_BATCH_SIZE_PER_GPU is defined in this format:
"train_micro_batch_size_per_gpu": 1
'''
TRAIN_MICRO_BATCH_SIZE_PER_GPU = "train_micro_batch_size_per_gpu"
TRAIN_MICRO_BATCH_SIZE_PER_GPU_DEFAULT = None
#########################################
# Gradient Accumulation
#########################################
# Gradient accumulation feature. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
GRADIENT_ACCUMULATION_FORMAT = '''
Gradient Accumulation should be of the format:
"gradient_accumulation_steps": 1
'''
GRADIENT_ACCUMULATION_STEPS = "gradient_accumulation_steps"
GRADIENT_ACCUMULATION_STEPS_DEFAULT = None
# DeepSpeed CSR gradient sparsity
SPARSE_GRADIENTS = "sparse_gradients"
SPARSE_GRADIENTS_DEFAULT = False
#########################################
# FP16 support
#########################################
# FP16 feature. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
FP16_FORMAT = '''
FP16 parameters should be of the format:
"fp16": {
"enabled": true,
"loss_scale": 0,
"initial_scale_power": 32,
"loss_scale_window": 1000,
"hysteresis": 2,
"min_loss_scale": 1
}
'''
FP16 = "fp16"
FP16_ENABLED = "enabled"
FP16_ENABLED_DEFAULT = False
# FP16 loss scale, zero means using dynamic scaling
FP16_LOSS_SCALE = "loss_scale"
FP16_LOSS_SCALE_DEFAULT = 0
# FP16 initial dynamic scale loss power
FP16_INITIAL_SCALE_POWER = "initial_scale_power"
FP16_INITIAL_SCALE_POWER_DEFAULT = 32
# FP16 loss scale window
FP16_LOSS_SCALE_WINDOW = "loss_scale_window"
FP16_LOSS_SCALE_WINDOW_DEFAULT = 1000
# FP16 hysteresis
FP16_HYSTERESIS = "hysteresis"
FP16_HYSTERESIS_DEFAULT = 2
# FP16 min loss scale
FP16_MIN_LOSS_SCALE = "min_loss_scale"
FP16_MIN_LOSS_SCALE_DEFAULT = 1
#########################################
# Apex AMP support
#########################################
# Use Apex AMP for mixed precision support, all parameters (other than 'enabled') will be passed to
# amp.initialize(model, optimizer, **amp_params)
# See apex documentation for supported parameters/features: https://nvidia.github.io/apex/amp.html#apex.amp.initialize
AMP_FORMAT = '''
"amp" {
"enabled: true,
"opt_level": "O1",
...
}
'''
AMP = "amp"
AMP_ENABLED = "enabled"
AMP_ENABLED_DEFAULT = False
#########################################
# Gradient clipping
#########################################
# Gradient clipping. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
GRADIENT_CLIPPING_FORMAT = '''
Gradient clipping should be enabled as:
"gradient_clipping": 1.0
'''
GRADIENT_CLIPPING = 'gradient_clipping'
GRADIENT_CLIPPING_DEFAULT = 0.
#########################################
# ZeRO optimization
#########################################
# ZeRO optimization. By default, this optimization is not enabled.
# Users have to configure the desired optimization (0 means disabled) in params.json as below example:
ZERO_FORMAT = '''
ZeRO optimization should be enabled as:
"session_params": {
"zero_optimization": [0|1|2],
"zero_all_gather_size": 200
}
'''
ZERO_OPTIMIZATION = 'zero_optimization'
ZERO_OPTIMIZATION_DEFAULT = 0
ZERO_OPTIMIZATION_OPTIMIZER_STATES = 1
ZERO_OPTIMIZATION_GRADIENTS = 2
ZERO_OPTIMIZATION_WEIGHTS = 3
MAX_STAGE_ZERO_OPTIMIZATION = ZERO_OPTIMIZATION_GRADIENTS
ZERO_REDUCE_SCATTER = "zero_reduce_scatter"
ZERO_REDUCE_SCATTER_DEFAULT = True
ZERO_MAX_ELEMENTS_PER_COMM = "zero_max_elements_per_comm"
ZERO_MAX_ELEMENTS_PER_COMM_DEFAULT = 5e8
ALLGATHER_SIZE = 'allgather_size'
ALLGATHER_SIZE_DEFAULT = 500000000
#########################################
# FP32 AllReduce
#########################################
# FP32 All reduce. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
FP32_ALLREDUCE_FORMAT = '''
FP32 Allreduce should be enabled as:
"fp32_allreduce": true
'''
FP32_ALLREDUCE = "fp32_allreduce"
FP32_ALLREDUCE_DEFAULT = False
#########################################
# Scale/predivide gradients before allreduce
#########################################
# Prescale gradients. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
PRESCALE_GRADIENTS_FORMAT = '''
Gradient prescaling should be enabled as:
"prescale_gradients": true
'''
PRESCALE_GRADIENTS = "prescale_gradients"
PRESCALE_GRADIENTS_DEFAULT = False
GRADIENT_PREDIVIDE_FACTOR_FORMAT = '''
Gradient predivide factor should be enabled as:
"gradient_predivide_factor": 1.0
'''
GRADIENT_PREDIVIDE_FACTOR = "gradient_predivide_factor"
GRADIENT_PREDIVIDE_FACTOR_DEFAULT = 1.0
#########################################
# Disable AllGather
#########################################
# Disable AllGather. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
DISABLE_ALLGATHER_FORMAT = '''
Disable AllGather should be enabled as:
"disable_allgather": true
'''
DISABLE_ALLGATHER = "disable_allgather"
DISABLE_ALLGATHER_DEFAULT = False
#########################################
# Dump DeepSpeed state
#########################################
# Dump State. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
DUMP_STATE_FORMAT = '''
Dump state should be enabled as:
"dump_state": true
'''
DUMP_STATE = 'dump_state'
DUMP_STATE_DEFAULT = False
#########################################
# Vocabulary size
#########################################
# Vocabulary size.
# Users can configure in ds_config.json as below example:
VOCABULARY_SIZE_FORMAT = '''
Vocabulary size can be specified as:
"vocabulary_size": 1024
'''
VOCABULARY_SIZE = 'vocabulary_size'
VOCABULARY_SIZE_DEFAULT = None
#########################################
# Wall block breakdown
#########################################
# Wall clock breakdown. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
WALL_CLOCK_BREAKDOWN_FORMAT = '''
Wall block breakdown should be enabled as:
"wall_clock_breakdown": true
'''
WALL_CLOCK_BREAKDOWN = 'wall_clock_breakdown'
WALL_CLOCK_BREAKDOWN_DEFAULT = False
MEMORY_BREAKDOWN = 'memory_breakdown'
MEMORY_BREAKDOWN_DEFAULT = False
#########################################
# Tensorboard
#########################################
# Tensorboard. By default, this feature is not enabled.
# Users can configure in ds_config.json as below example:
TENSORBOARD_FORMAT = '''
Tensorboard can be specified as:
"tensorboard": {
"enabled": true,
"output_path": "/home/myname/foo",
"job_name": "model_lr2e-5_epoch3_seed2_seq64"
}
'''
TENSORBOARD = "tensorboard"
# Tensorboard enable signal
TENSORBOARD_ENABLED = "enabled"
TENSORBOARD_ENABLED_DEFAULT = False
# Tensorboard output path
TENSORBOARD_OUTPUT_PATH = "output_path"
TENSORBOARD_OUTPUT_PATH_DEFAULT = ""
# Tensorboard job name
TENSORBOARD_JOB_NAME = "job_name"
TENSORBOARD_JOB_NAME_DEFAULT = "DeepSpeedJobName"
| [
"noreply@github.com"
] | noreply@github.com |
f9184c9bc8feb760a6b06e8e4065cacd924722c5 | 24223cba74ca8c293aaa4381fecf80acd237fc94 | /jensen_comparison/bin/jupyter-kernelspec | 048319db09ad921800f7139bc93ab93c2745267a | [] | no_license | Aequivinius/jensen-comparison | df0f4424160d5c4d37f29eb8939518c96b446b9c | 04910a40869103a618dd4cea5035ff2c9c6797bf | refs/heads/master | 2020-04-30T00:44:18.755131 | 2019-03-19T12:48:10 | 2019-03-19T12:48:10 | 176,510,273 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 312 | #!/Users/sazerac/jensen_comparison/jensen_comparison/bin/python3.7
# -*- coding: utf-8 -*-
import re
import sys
from jupyter_client.kernelspecapp import KernelSpecApp
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(KernelSpecApp.launch_instance())
| [
"ncolic@gmail.com"
] | ncolic@gmail.com | |
7d0942492c486ab43f4c39a5adee4453c034f50e | c1fe97208afe479b7ae1ee67d69866a6911564ca | /AdvCBV/basicapp/admin.py | 046528c9de692ae1d3f199606430ad8437e9c4a1 | [] | no_license | jaindhairyahere/Python_Django | a0a46c57b6ca60d0942ae181fe28ea56bb1ee948 | f170a2e38b78df698a02821a454a3baea0c358a6 | refs/heads/master | 2020-06-18T09:17:56.364928 | 2019-11-02T18:34:12 | 2019-11-02T18:34:12 | 196,249,108 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 164 | py | from django.contrib import admin
from basicapp.models import School, Student
# Register your models here.
admin.site.register(Student)
admin.site.register(School)
| [
"jaindhairya2001@gmail.com"
] | jaindhairya2001@gmail.com |
151fc23e1533e76eb12ce1b8bb1392755667dbab | 7f54637e347e5773dfbfded7b46b58b50544cfe5 | /7-3/chainxy/spiders/tradesecretscanada.py | dc8b30b2b1997267ec5b41a42628814c788f3cc0 | [] | no_license | simba999/all-scrapy | 5cc26fd92b1d03366b74d4fff58c4a0641c85609 | d48aeb3c00fa2474153fbc8d131cf58402976e1d | refs/heads/master | 2021-01-25T14:24:04.715550 | 2018-03-03T13:43:13 | 2018-03-03T13:43:13 | 123,695,640 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,369 | py | import scrapy
import json
import os
from scrapy.spiders import Spider
from scrapy.http import FormRequest
from scrapy.http import Request
from chainxy.items import ChainItem
from lxml import etree
from selenium import webdriver
from lxml import html
import usaddress
import pdb
class tradesecretscanada(scrapy.Spider):
name = 'tradesecretscanada'
domain = ''
history = []
def start_requests(self):
init_url = 'http://talk.tradesecrets.ca/locations-reviews/'
yield scrapy.Request(url=init_url, callback=self.body)
def body(self, response):
print("========= Checking.......")
store_list = response.xpath('//a[@rel="noopener noreferrer"]/@href').extract()
for store in store_list:
yield scrapy.Request(url=store, callback=self.parse_page)
def parse_page(self, response):
try:
item = ChainItem()
detail = self.eliminate_space(response.xpath('//div[contains(@class, "fusion-one-half fusion-layout-column fusion-spacing-no")]//h4//text()').extract())
h_temp = ''
for de in detail:
if '(' in de and '-' in de:
try:
item['phone_number'] = self.validate('(' + de.split('(')[1])
except:
item['phone_number'] = self.validate(de)
if ':' in de:
h_temp += de + ', '
if '(' in detail[0]:
detail[0] = self.validate(detail[0].split('(')[0]).replace('|','')
addr = detail[0].replace('|','').split(',')
if len(addr) == 4:
item['address'] = self.validate(addr[1])
item['city'] = self.validate(addr[2])
item['state'] = self.validate(addr[3].strip())[:2].strip()
item['zip_code'] = self.validate(addr[3])[2:].strip()
elif len(addr) == 3:
item['address'] = self.validate(addr[0])
item['city'] = self.validate(addr[1])
item['state'] = self.validate(addr[2].strip())[:2].strip()
item['zip_code'] = self.validate(addr[2])[2:].strip()
else:
pdb.set_trace()
item['country'] = 'Canada'
item['store_hours'] = h_temp[:-2]
yield item
except:
pass
def validate(self, item):
try:
return item.encode('raw-unicode-escape').replace('\u2013', '').replace('\xa0', '').replace('|','').strip()
except:
return ''
def eliminate_space(self, items):
tmp = []
for item in items:
if self.validate(item) != '' and 'try' not in self.validate(item).lower() and 'http' not in self.validate(item).lower():
tmp.append(self.validate(item))
return tmp | [
"oliverking8985@yahoo.com"
] | oliverking8985@yahoo.com |
16609f41774f606706d680afcec9bb61b9b40b3d | e540673e49a0ea897b0df0d15c2c5df38e54e8b0 | /TF_polynomial_regression_2.py | 85025c0668bb062ec899d2646b6dc3697620fdbc | [] | no_license | rottendoom001/PYTHON_TF | ec97a287ec71cfb24e4bf94cc27df36c8d451a1a | 0de5489b08f60c91bb71145c16836bbf91174409 | refs/heads/master | 2021-09-03T17:56:04.114074 | 2018-01-10T22:06:01 | 2018-01-10T22:06:01 | 109,761,949 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,986 | py | """Simple tutorial for using TensorFlow to compute polynomial regression.
Parag K. Mital, Jan. 2016"""
# %% Imports
import numpy as np
import tensorflow as tf
import matplotlib.pyplot as plt
# Normalizar datos
def normalize (xs):
ln = []
maxValue = max(xs)
minValue = min(xs)
for e in xs :
# Formula para nomalizar
# (x - min(x)) / (max(x) - min(x))
r = ( e - minValue ) / ( maxValue - minValue )
ln.append(r)
return ln
# Construir el vector de valores para X
def build_x_vector(x, grade):
x_new = np.zeros([x.size, grade])
for i in range(grade):
x_new[:,i] = np.power(x,(i+1))
return x_new
# %% Let's create some toy data
plt.ion()
fig, ax = plt.subplots(1, 1)
#xs = np.linspace(-3, 3, n_observations)
#ys = np.sin(xs) + np.random.uniform(-0.5, 0.5, n_observations)
xs = [1,2,3,4,5,6,7,8,9,10,11,12,13]
xs = np.array(xs)
#ys = [449,525,412,161,639,732]
ysn = [15, 14, 13, 6, 19, 16, 16, 11, 17, 15, 12, 25, 14]
#ysn = normalize(ys)
print (ysn)
ax.scatter(xs, ysn)
fig.show()
plt.draw()
# Grado de la ecuación polinomial
n = 15
W = tf.Variable(tf.random_normal([n,1]), name='weight')
b = tf.Variable(tf.random_normal([1]), name='bias')
X = tf.placeholder(tf.float32)
Y = tf.placeholder(tf.float32)
X_new = build_x_vector(xs, n)
print("X_new : \n%s"%(X_new))
Y_pred = tf.add(tf.matmul(X,W),b)
loss = tf.reduce_mean(tf.square(Y_pred - Y))
learning_rate = 0.01
optimizer = tf.train.GradientDescentOptimizer(learning_rate).minimize(loss)
n_epochs = 1000
with tf.Session() as sess:
for step in range(n_epochs):
sess.run(tf.global_variables_initializer())
_, c = sess.run([optimizer, loss], feed_dict={X: X_new, Y: ysn})
if step % 10 == 0:
print("COSTO: %s"%(c))
print ("W: %s"%(sess.run(W)))
print ("b: %s"%(sess.run(b)))
y_test = sess.run(Y_pred, feed_dict={X: X_new})
ax.scatter(xs, y_test)
fig.show()
plt.draw()
plt.waitforbuttonpress()
| [
"alanorlando.cruz.manrique@bbva.com"
] | alanorlando.cruz.manrique@bbva.com |
8fe101007d836f2e2cedb46604f7ca5e40fe18f5 | 3b8deb008d8bb5199e4ea37a0715886bf02559c0 | /examples/fakerstc.py | 0f4eaa7793229cdf198523acd71dec790b001314 | [] | no_license | ryanmwhitephd/ADD | 759be1b8afd4433ff9d76606ad3deead64814977 | 60ad72f0e4360cf66e204451efbc9a5baa0732fa | refs/heads/master | 2021-09-18T14:40:17.398413 | 2018-07-16T06:57:01 | 2018-07-16T06:57:01 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,834 | py | # -*- coding: utf-8 -*-
"""
Created on Thu Jan 25 12:24:32 2018
@author: ryanwhi
"""
import logging
import unittest
import argparse
import json
import functools
log = logging.getLogger(__name__)
log.addHandler(logging.NullHandler())
from faker.providers import BaseProvider
from faker import Faker
# Use functools.partial
# https://stackoverflow.com/questions/29832410/python-use-dictionary-keys-as-function-names
class ModelType(object):
'''
Base Model class
provides all the required functions
needed for accessing properties
of the model
'''
def __init__(self,mdata):
self.__meta = mdata
@property
def meta(self):
return self.__meta
@meta.setter
def meta(self,mdata):
self.__meta = mdata
def provider(self,name):
return self.meta[name]
class MetaModel(object):
'''
Wrapper class to instatiate
user-defined model
provides easy access to
Faker providers and optional arguments
'''
def __init__(self,name):
self.model = globals()[name]
@property
def model(self):
return self.__model
@model.setter
def model(self,ctor):
self.__model = ctor()
def providers(self):
'''
return model dict
'''
return self.model.meta
class MyModel(ModelType):
'''
User-defined model
Define a dictionary
with the Faker provider (or custom provider)
optional arguments
@TODO
add the header information for column names
'''
def __init__(self):
meta = {'name':None,
'ean':13}
super(MyModel,self).__init__(meta)
class AltModel(ModelType):
def __init__(self):
meta = {'name':None,
'address':None,
'longitude':None,
'iban':None,
'credit_card_full':'visa16'
}
super(AltModel,self).__init__(meta)
def fakerstc(*args):
mymodel = MetaModel(args[0])
print(mymodel.model.meta)
print(mymodel.providers())
print(mymodel.model.provider('name'))
fake = Faker('en_CA')
for _ in range(args[1]):
row=[]
for key in mymodel.providers():
parms=mymodel.providers()[key]
if parms is None:
row.append(functools.partial(fake.__getattribute__(key))())
else:
row.append(functools.partial(fake.__getattribute__(key))(parms))
print(row)
if __name__ == '__main__':
fakerstc('MyModel',10)
fakerstc('AltModel',10) | [
"noreply@github.com"
] | noreply@github.com |
f5b01891fa5b5a79a702ba2de7a0f3e064675000 | 6d86d4e07eeed66fdc08f68560dce9696bb35c01 | /app/main/errors.py | d787733609d6910457a6556500d9d042d3fa9ec0 | [] | no_license | Moxikai/law_show | 5d4828a85a2bf1ae530a7e1f3d87e117d2fec4aa | 9819b263a7c22680695d8a04a3992a627a73f585 | refs/heads/master | 2021-01-13T10:30:58.666067 | 2016-10-13T08:47:31 | 2016-10-13T08:47:31 | 68,830,448 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 323 | py | #_*_coding:utf-8_*_
from flask import render_template
from . import main
#使用app_errorhandler修饰器,处理全局错误
@main.app_errorhandler(404)
def page_not_found(e):
return render_template('404.html'),404
@main.app_errorhandler(500)
def internal_server_error(e):
return render_template('500.html'),500
| [
"zhu-hero@qq.com"
] | zhu-hero@qq.com |
7548f99c00f871df535a5e2a2a86c932211f256d | 3609f3da79af9ad44f047c58edd92da590afa17e | /PythonExerciseFolder/WearwolfOutTheDoor.py | 11961bf0da3d0167ce6cc75f3e556955080e28fc | [] | no_license | Poludzku/master | bcfaab3d4e8e20d4cb5b6e31ea349342c67a7f8f | 8fbc44b30b35f9b0e7608782da4d61e673d68bab | refs/heads/master | 2020-12-25T18:22:24.650293 | 2012-07-28T12:25:13 | 2012-07-28T12:25:13 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 905 | py | # -*- coding: utf-8 *-*
from WearwolfZones import mountines
from WearwolfZones import spring_oasis
from WearwolfZones import forest
def steping_out_the_door():
print """As you take a step out the door the fresh air fills your lungs,
you can feel that tonight it will be a good kill, but where will you hunt?
In the mountines which Anuk calls his home and the moste elite predators
reside, the well spring oasis where many fine elks make there wattering hole
or the forest benethe the mountines where old hunters claim they have seen
the mystical blue fox"""
next = raw_input('> ')
if next == "mountines":
mountines()
elif next == "spring oasis":
spring_oasis()
elif next == "forest":
forest()
else:
print """You can go hunt in the 'mountines', the great forest or the
spring_oasis"""
steping_out_the_door()
| [
"jacek@mint"
] | jacek@mint |
2370e7452bcc9e77a37e5853184a510e1184341d | d2c4934325f5ddd567963e7bd2bdc0673f92bc40 | /tests/model_control/detailed/transf_Logit/model_control_one_enabled_Logit_MovingMedian_NoCycle_LSTM.py | 52f0b5d941ab48694348dbd8ae8a86fd89845917 | [
"BSD-3-Clause",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | jmabry/pyaf | 797acdd585842474ff4ae1d9db5606877252d9b8 | afbc15a851a2445a7824bf255af612dc429265af | refs/heads/master | 2020-03-20T02:14:12.597970 | 2018-12-17T22:08:11 | 2018-12-17T22:08:11 | 137,104,552 | 0 | 0 | BSD-3-Clause | 2018-12-17T22:08:12 | 2018-06-12T17:15:43 | Python | UTF-8 | Python | false | false | 155 | py | import pyaf.tests.model_control.test_ozone_custom_models_enabled as testmod
testmod.build_model( ['Logit'] , ['MovingMedian'] , ['NoCycle'] , ['LSTM'] ); | [
"antoine.carme@laposte.net"
] | antoine.carme@laposte.net |
be6628d9dba373743c57c916d5222277a1ceab7e | e4284c03268b662eb6d121fedd32aedbabd2ae4b | /test/FileIoTest.py | 59794fc60bae2af560a73de5df088c4f42da0105 | [] | no_license | rennX/FunctionPointParser | 99af5588158163f78b07dd3b4bfb8196db71b557 | 65fce373fae5baf1b34e8a2863c7358981139fa6 | refs/heads/master | 2016-09-09T23:42:56.676573 | 2014-04-20T14:02:36 | 2014-04-20T14:02:36 | 16,235,300 | 1 | 0 | null | 2014-01-26T22:53:35 | 2014-01-25T17:05:10 | null | UTF-8 | Python | false | false | 474 | py | #! /usr/bin/env python
"""This class reads in and processes
input text from a file"""
import unittest
class FileIoTest(unittest.TestCase):
det setUp(self):
self.fileIo = FileIo("input.txt")
def tearDown(self):
self.fileIo = None
def toStringTest(self):
"""Prints out the contents of the input file"""
assert
print str(self.fileIo.inputFile())
def getFileTest(self):
"""Returns the contents of the input file as a string"""
return self.inputFile
| [
"renn.xa@gmail.com"
] | renn.xa@gmail.com |
efafbc94b23a7a62be62c4ca61b6578234ec47c0 | 92e2145b2424134112ef0d7b1b5cf1c1220dbc1e | /Sith/wsgi.py | fe015aa0f796905984ddf362e66f7fe3c8aba921 | [] | no_license | rustam66637/SithRecruitTest | 9c52725d99df8ae074068d17e7412565a12d9f2d | 1f8fd5da3b71d8b1af59b6d8214807ec37941dc5 | refs/heads/master | 2023-08-06T17:02:51.966481 | 2021-02-07T15:48:20 | 2021-02-07T15:48:20 | 265,617,598 | 0 | 0 | null | 2021-09-22T19:03:30 | 2020-05-20T16:04:42 | Python | UTF-8 | Python | false | false | 385 | py | """
WSGI config for Sith project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/3.0/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'Sith.settings')
application = get_wsgi_application()
| [
"rust66637@gmail.com"
] | rust66637@gmail.com |
a8b4064eeb4bb83cbf8897edf30fd9ef90b99913 | 88940cbee65e15f8d9dce61ea313a03d38131e85 | /fbproj.py | 59263db4e13a12a883d849aa33f66cb8d17624a3 | [] | no_license | hanslee95/206FB | 025fd961deffbab868238addd276177cc3839e5a | f79640b84c9f47a75b98cadded6228ec0784afac | refs/heads/master | 2021-08-28T21:38:44.846473 | 2017-12-13T06:44:25 | 2017-12-13T06:44:25 | 112,550,535 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,275 | py | import urllib3
import facebook
import requests
import json
import datetime
import sqlite3
import plotly.plotly as py
import plotly.graph_objs as go
import pandas as pd
from facepy import GraphAPI
#NOTE: I have gotten 100 interactions extracting 100 posts, however, the final data table will have 7 items (rows) in each column.
#as it's information on the 7 days of the week. Specifically, in each post of the 100, I collected how many times I either posted a story or message to see how active I am on the certain day.
# ############## CACHING THE DATA ###########################################################################################################################################################
access_token = 'access_token'
graph = GraphAPI(access_token)
# ############## CACHING THE DATA ###########################################################################################################################################################
CACHE_FNAME = "Posts_cache.json"
try:
cache_file = open(CACHE_FNAME, 'r') # Try to read the data from the file
cache_contents = cache_file.read() # If it's there, get it into a string
CACHE_DICTION = json.loads(cache_contents) # And then load it into a dictionary
cache_file.close() # Close the file, we're good, we got the data in a dictionary.
except:
CACHE_DICTION = {}
#checking the cache of a particular user and returns that data or retrieves that cache'd data.
def get_posts():
# if statement is checking if you already looked it up, if you did, then use the thing you cache'd
if 'posts' in CACHE_DICTION:
print('using cache')
post_results = CACHE_DICTION['posts']
else:
print('getting data from internet')
post_results = graph.get('me?fields=posts.limit(118)')
CACHE_DICTION['posts'] = post_results['posts']
cache_file = open(CACHE_FNAME, 'w')
#json.dumps prints out the string of dictionary in a json file in one line
cache_file.write(json.dumps(CACHE_DICTION))
cache_file.close()
return CACHE_DICTION['posts']
posts = get_posts()
########## HELPER FUNCTIONS #####################################################################################################################
#To convert the datetime to a day of the week
def get_day_of_week(date):
return datetime.datetime.strptime(date, '%Y-%m-%d').strftime('%a')
#stripping the created_time value to only get the year-month-day
def strip_time(time_str):
date = time_str[:-14]
return date
#iterate over the weeks to compare when each element was posted. Then iterate over the dictionary and if that post happened on
#first day (mon), then check and add the stories and messages separately. Then Repeat. Returns dictionary with week day as keys and
#list of tuples (activites)
def calculate_activity(week_l, p):
week_dict = {}
for day in week_l:
for elem in p['data']:
m_count = 0
s_count = 0
#if the created time is equal to that day of the week. Only getting message/story from that day
if get_day_of_week(strip_time(elem['created_time'])) == day:
#if that day has not been added to the dictionary, add it
if get_day_of_week(strip_time(elem['created_time'])) not in week_dict:
week_dict[get_day_of_week(strip_time(elem['created_time']))] = []
else:
for key in elem:
if key != 'story' and key == 'message':
m_count = m_count + 1
if key == 'story' and key != 'message':
s_count = s_count + 1
# else both story and message are in that date so add two activities
if key == 'story' and key == 'message':
m_count = m_count + 1
s_count = s_count + 1
tup = s_count, m_count
week_dict[get_day_of_week(strip_time(elem['created_time']))].append(tup)
return week_dict
#adding up all the elements in list of tuples for story and message then creating 3 element tuple to insert to table.
def insert_tup3(d):
for key in d:
add_s = sum(i[0] for i in d[key])
add_m = sum(i[1] for i in d[key])
tup = add_s, add_m
d[key] = []
d[key].append(tup)
for x in d[key]:
tup3 = key, x[0], x[1]
cur.execute('INSERT INTO WeekDay (created_time, story, message) VALUES (?, ?, ?)', tup3)
conn.commit()
########### CREATING AND LOADING IN DATA INTO DATABASE #################################################################################################
conn = sqlite3.connect('Posts_Day.sqlite', timeout = 10)
cur = conn.cursor()
cur.execute('DROP TABLE IF EXISTS WeekDay')
cur.execute("CREATE TABLE WeekDay (created_time TEXT, story NUMBER, message NUMBER)")
week_lst = ['Mon', 'Tue', 'Wed', 'Thu', 'Fri', 'Sat', 'Sun']
#calling helper functions to insert data in database
insert_tup3(calculate_activity(week_lst, posts))
#getting data for x-axis
cur.execute('SELECT created_time FROM WeekDay')
lst_x = cur.fetchall()
lst_day = [elem[0] for elem in lst_x]
#getting data for y-axis (story)
cur.execute('SELECT story FROM WeekDay')
lst_y1 = cur.fetchall()
lst_activity = [elem[0] for elem in lst_y1]
#getting data for y-axis (message)
cur.execute('SELECT message FROM WeekDay')
lst_y2 = cur.fetchall()
lst_message = [elem[0] for elem in lst_y2]
########### CREATE DATA VIZ USING PLOTLY #################################################################################################
#importing the data to create graph
trace_high = go.Scatter(
x=lst_day,
y=lst_activity,
name = "Activity",
line = dict(color = '#03B9DF'),
opacity = 0.8)
trace_low = go.Scatter(
x=lst_day,
y=lst_message,
name = "Message",
line = dict(color = '#00E264'),
opacity = 0.8)
data = [trace_high,trace_low]
#labeling graph with title, x/y axis.
layout = dict(
title = "FACEBOOK LIFE",
xaxis=dict(
title='Day of Week',
titlefont=dict(
family='Courier New, monospace',
size=18,
color='#7f7f7f'
)
),
yaxis=dict(
title='Number of Posts',
titlefont=dict(
family='Courier New, monospace',
size=18,
color='#7f7f7f'
)
)
)
fig = dict(data=data, layout=layout)
py.iplot(fig, filename = "FACEBOOK LIFE")
| [
"hanheum@umich.edu"
] | hanheum@umich.edu |
62f1f7e28e890ada8f842b1295e4c295bd93ce6b | 3d14e2430b696a21661ee1b5fc000aa031a81a67 | /locators/books_page_locators.py | 403e152b6985de415825fb662971e8422cf916ea | [] | no_license | a-soliman/books-toscrap | 53defb67df7167917a53fa1193fff5f3bccd998c | 958c347eeb881178346f228326b4ccfe47b7acd2 | refs/heads/master | 2020-04-25T17:25:36.492827 | 2019-02-28T16:22:59 | 2019-02-28T16:22:59 | 172,947,801 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 114 | py | """
extracts books from a page
"""
class BooksPageLocators:
BOOK = "li.col-xs-6.col-sm-4.col-md-3.col-lg-3"
| [
"ahmed.soliman@programmer.net"
] | ahmed.soliman@programmer.net |
b119a388c312e6aa8b39e003c88c95a830239c8a | 39e1f30888267253a0e2976c05e364c001f386df | /validate-user.py | 81499137b2e28d5885e4b4ca891499f25dc7672b | [] | no_license | atomWeb/facegnition-back-end | e1ce12aced40bc9fc902fa3d86695746f811814e | f6ef3ec256c05f5db64d25e12c006aef12f5848d | refs/heads/main | 2023-08-17T01:05:32.200992 | 2021-10-08T10:32:33 | 2021-10-08T10:32:33 | 403,296,567 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,631 | py | import json
import boto3
from boto3.dynamodb.conditions import Key
import os
from utils import jsonify, get_str_timestamp_by_zone
S3_BUCKET = os.environ["IMAGE_BUCKET"]
REGION = os.environ["REGION"]
USERS_TABLE = os.environ["USERS_TABLE"]
VALIDATIONS_TABLE = os.environ["VALIDATIONS_TABLE"]
COLLECT_NAME = os.environ["REKOGN_COLLECTION"]
ZONE = os.environ["TIME_ZONE"]
s3_client = boto3.client("s3", region_name=REGION)
dynamo_resource = boto3.resource("dynamodb", region_name=REGION)
users_table = dynamo_resource.Table(USERS_TABLE)
validations_table = dynamo_resource.Table(VALIDATIONS_TABLE)
rekog = boto3.client('rekognition', region_name=REGION)
def validate(event, context):
print(event)
jresp = {'data': ''}
statusCode=200
faceId = "None"
usr_name = ""
try:
# Datos de la ejecución
data = json.loads(event["body"])
s3path = data["imgname"]
# buscar caras por imagen en una colección
response = rekog.search_faces_by_image(
CollectionId=COLLECT_NAME,
Image={
'S3Object': {
'Bucket': S3_BUCKET,
'Name': s3path
}
},
MaxFaces=5,
FaceMatchThreshold=97.0,
QualityFilter='AUTO'
)
print("Response rekognition buscar cara: ", response)
if not response['FaceMatches']:
statusCode = 404
else:
faceId = response['FaceMatches'][0]['Face']['FaceId']
timestamp = get_str_timestamp_by_zone(ZONE)
# Guarda en la DB el registro de la validacion.
response = validations_table.put_item(
Item={
'faceid': faceId,
'timest': timestamp,
's3path': s3path
}
)
print("Response Dynamo tabla validations: ", response)
# Nota: Al guardar este registro en dynamo disparará la lambda que verifica el notify en la tabla de user y lanzara el email
usrs_resp = users_table.query(
KeyConditionExpression=Key("faceid").eq(faceId),
Limit=1
)
items = usrs_resp.get('Items', [])
if items:
usr_name = items[0]["name"]
except Exception as e:
msgError = "An exception occurred " + str(e) + "."
print(msgError)
jresp = {'Error': msgError}
statusCode = 500
jresp = {"face-id": faceId, "name": usr_name}
return jsonify(jresp, statusCode)
| [
"cristianjose@gmail.com"
] | cristianjose@gmail.com |
e3a17a5121943d355e6b723092ef64da003de344 | 40b52661769a803f470b84872d692a388095cca6 | /lovelace/migrations/0013_auto_20190319_2002.py | 67687e3fc0868d8e75c8be7e5a20dd96cdf83b91 | [] | no_license | leilesca/demodayfinalizado | 4ff2e3fc4fabfa85c96cd4778b257cb20d00dd71 | edb117f3d73bcda53f6db0625335e3bbd688f6f8 | refs/heads/master | 2020-04-30T06:55:57.077625 | 2019-03-20T15:14:00 | 2019-03-20T15:14:00 | 176,668,242 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 582 | py | # Generated by Django 2.1.7 on 2019-03-19 23:02
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('lovelace', '0012_perfilusuario'),
]
operations = [
migrations.AddField(
model_name='estabelecimento',
name='imagem',
field=models.CharField(default='', max_length=250),
),
migrations.AlterField(
model_name='estabelecimento',
name='telefone_estabelecimento',
field=models.CharField(max_length=20),
),
]
| [
"alessandra.mitie@gmail.com"
] | alessandra.mitie@gmail.com |
92e1f7036fa5a7c131b843407e94744c4c5e5dee | 67212da14d34fe69de64dcc14301a6d3f2004b89 | /main.py | 95789d840eaf50d28773ee383491fa71b7346331 | [] | no_license | seiya-tsukada/my_gae_sample | 7212b60bc48da4a06a7ff3d83c2c0a1db4dbc28f | 4223923fb00c2b75fac0520f5938bfdeb9b3db92 | refs/heads/master | 2020-04-01T23:45:36.970914 | 2018-10-19T15:47:17 | 2018-10-19T15:47:17 | 153,772,079 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 250 | py | #! /usr/bin/env python
# coding: utf-8
from flask import Flask, jsonify
app = Flask(__name__)
app.config['JSON_AS_ASCII'] = False
@app.route("/")
def index():
return jsonify({"message": "hello world"})
if __name__ == "__main__":
app.run() | [
"seiya.tsukada@accenture.com"
] | seiya.tsukada@accenture.com |
5c6e5146bb6f11b1d58dd88282e817ec96737c23 | 6051b1df3a84ca6dc3d4c4ce90202e64a31da22b | /test_studentlists.py | 8d714b81118ded8f1a951b467ba98d2ec7583061 | [] | no_license | omarmohamud23/Labweek4--studentlist | e4c614ee93c0cf2e911210aaeeb368ef37916953 | 8b11a661ffdc226c049d48b4d4e0479926e004a4 | refs/heads/master | 2023-02-26T22:32:06.700044 | 2021-02-05T21:59:53 | 2021-02-05T21:59:53 | 336,397,878 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,545 | py | '''
Practice using
assertTrue
assertFalse
assertIsNone
assertIsNotNone
assertIn
assertNotIn
'''
from studentlists import ClassList, StudentError
from unittest import TestCase
class TestStudentLists(TestCase):
def test_add_student_check_student_in_list(self):
test_class = ClassList(2)
test_class.add_student('Test Student')
self.assertIn('Test Student', test_class.class_list)
test_class.add_student('Another Test Student')
self.assertIn('Test Student', test_class.class_list)
self.assertIn('Another Test Student', test_class.class_list)
def test_add_student_already_in_list(self):
test_class = ClassList(2)
test_class.add_student('Test Student')
with self.assertRaises(StudentError):
test_class.add_student('Test Student')
## TODO write a test that adds and removes a student,
# and asserts the student is removed. Use assertNotIn
def test_add_remove_student_ensure_removed(self):
test_class = ClassList(2)
test_class.add_student('Test Student')
test_class.remove_student('Test Student')
self.assertNotIn('Test Student', test_class.class_list)
## TODO write a test that adds some example students,
# then removes a student not in the list, and asserts a StudentError is raised
def test_add_students_remove_student_not_in_list(self):
test_class = ClassList(4)
test_class.add_student('Student One')
test_class.add_student('Student Two')
test_class.add_student('Student Three')
with self.assertRaises(StudentError):
test_class.remove_student('Student Four')
## TODO write a test that removes a student from an
# empty list, and asserts a StudentError is raised
def test_removes_student_in_empty_list(self):
test_class = ClassList(3)
with self.assertRaises(StudentError):
test_class.remove_student('Abdi')
def test_is_enrolled_when_student_present(self):
test_class = ClassList(2)
test_class.add_student('Snoop Dogg')
test_class.add_student('Martha Stewart')
self.assertTrue(test_class.is_enrolled('Snoop Dogg'))
self.assertTrue(test_class.is_enrolled('Martha Stewart'))
def test_is_enrolled_empty_class_list(self):
test_class = ClassList(2)
self.assertFalse(test_class.is_enrolled('Snoop Dogg'))
## TODO write a test that adds some example students to a test class,
## then, call is_enrolled for a student who is not enrolled.
# Use assertFalse to verify is_enrolled returns False.
def add_student_examples_check_student_not_enrolled(self):
test_class = ClassList(2)
test_class.add_student('Alhandro')
test_class.add_student('Mutina')
self.assertFalse(test_class.is_enrolled('Abdi'))
def test_string_with_students_enrolled(self):
test_class = ClassList(2)
test_class.add_student('Taylor Swift')
test_class.add_student('Kanye West')
self.assertEqual('Taylor Swift, Kanye West', str(test_class))
def test_string_empty_class(self):
test_class = ClassList(2)
self.assertEqual('', str(test_class))
def test_index_of_student_student_present(self):
test_class = ClassList(3)
test_class.add_student('Harry')
test_class.add_student('Hermione')
test_class.add_student('Ron')
self.assertEqual(1, test_class.index_of_student('Harry'))
self.assertEqual(2, test_class.index_of_student('Hermione'))
self.assertEqual(3, test_class.index_of_student('Ron'))
# This assert passes, but it's redundant - the first assert statement will fail if
# the method call returns None
self.assertIsNotNone(test_class.index_of_student('Harry'))
## TODO write a test for index_of_student when the class_list list is empty.
# Assert index_of_student returns None for a student if the list is empty.
# use assertIsNone.
def test_index_of_student_in_empty_list_return_None(self):
test_class = ClassList(1)
self.assertIsNone(test_class.index_of_student('Hamze'))
## TODO write another test for index_of_student. In the case when the
# class_list is not empty but has some students.
# assert that searching for a student name that is not in the list, returns None.
def test_index_of_student_list_not_empty_search_student_not_in_list(self):
test_class = ClassList(2)
test_class.add_student('Mohamud')
test_class.add_student('Muhumed')
self.assertIsNone(test_class.index_of_student('Mary'))
## TODO write a test for your new is_class_full method when the class is full.
# use assertTrue.
def test_is_class_full(self):
test_class = ClassList(4)
test_class.add_student('Farmaajo')
test_class.add_student('Mukami')
test_class.add_student('Abdi')
test_class.add_student('Bisharo')
self.assertTrue(test_class.is_class_full())
## TODO write a test for your new is_class_full method for when is empty,
# and when it is not full. Use assertFalse.
def test_is_class_full_empty(self):
test_class = ClassList(2)
self.assertFalse(test_class.is_class_full())
def test_is_class_full_when_not_full(self):
test_class = ClassList(4)
test_class.add_student('Otieno')
test_class.add_student('odinga')
self.assertFalse(test_class.is_class_full()) | [
"Omar"
] | Omar |
b7c679595c429b8800f1d4f502e2a7f308771f09 | c478f4c1c14a250633666b9eac9835b36f86f704 | /ckcore/core/task/task_handler.py | dbfabe44bf93545706eaeb0b91447a503a52210a | [
"Apache-2.0"
] | permissive | yasinai/cloudkeeper | 10d49df5e9a69b4106ca16fff6bb04e150dd743d | 7d1708deaf237b4175006852d87a2ecce1372779 | refs/heads/main | 2023-08-14T07:10:33.071714 | 2021-10-13T08:37:23 | 2021-10-13T08:37:23 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 26,786 | py | from __future__ import annotations
import asyncio
import logging
import re
from asyncio import Task, CancelledError
from datetime import timedelta
from io import TextIOWrapper
from typing import Optional, Any, Callable, Union, Sequence, Dict, List, Tuple
import argparse
from aiostream import stream
from argparse import ArgumentParser, Namespace
from functools import reduce
from copy import copy
from core.cli.cli import CLI
from core.db.jobdb import JobDb
from core.db.runningtaskdb import RunningTaskData, RunningTaskDb
from core.error import ParseError, CLIParseError
from core.message_bus import MessageBus, Event, Action, ActionDone, Message, ActionError
from core.task.job_handler import JobHandler
from core.task.model import Subscriber
from core.task.scheduler import Scheduler
from core.task.start_workflow_on_first_subscriber import wait_and_start
from core.task.subscribers import SubscriptionHandler
from core.task.task_description import (
Workflow,
RunningTask,
EventTrigger,
TimeTrigger,
TaskSurpassBehaviour,
PerformAction,
Step,
TaskDescription,
Job,
ExecuteCommand,
TaskCommand,
SendMessage,
ExecuteOnCLI,
StepErrorBehaviour,
RestartAgainStepAction,
)
from core.util import first, Periodic, group_by, uuid_str, utc_str
log = logging.getLogger(__name__)
class TaskHandler(JobHandler):
# region init
@staticmethod
def add_args(arg_parser: ArgumentParser) -> None:
arg_parser.add_argument(
"--jobs",
nargs="*",
type=argparse.FileType("r"),
help="Read job definitions from given file.",
)
arg_parser.add_argument(
"--start-collect-on-subscriber-connect",
default=False,
action="store_true",
help="Start the collect workflow, when the first handling actor connects to the system.",
)
def __init__(
self,
running_task_db: RunningTaskDb,
job_db: JobDb,
message_bus: MessageBus,
subscription_handler: SubscriptionHandler,
scheduler: Scheduler,
cli: CLI,
args: Namespace,
):
self.running_task_db = running_task_db
self.job_db = job_db
self.message_bus = message_bus
self.subscription_handler = subscription_handler
self.scheduler = scheduler
self.cli = cli
self.args = args
# Step1: define all workflows and jobs in code: later it will be persisted and read from database
self.task_descriptions: Sequence[TaskDescription] = [*self.known_workflows(), *self.known_jobs()]
self.tasks: Dict[str, RunningTask] = {}
self.message_bus_watcher: Optional[Task] = None # type: ignore # pypy
self.initial_start_workflow_task: Optional[Task] = None # type: ignore # pypy
self.timeout_watcher = Periodic("task_timeout_watcher", self.check_overdue_tasks, timedelta(seconds=10))
self.registered_event_trigger: List[Tuple[EventTrigger, TaskDescription]] = []
self.registered_event_trigger_by_message_type: Dict[str, List[Tuple[EventTrigger, TaskDescription]]] = {}
# endregion
# region startup and teardown
async def update_trigger(self, desc: TaskDescription, register: bool = True) -> None:
# safeguard: unregister all event trigger of this task
for existing in (tup for tup in self.registered_event_trigger if desc.id == tup[1].id):
self.registered_event_trigger.remove(existing)
# safeguard: unregister all schedule trigger of this task
for job in self.scheduler.list_jobs():
if str(job.id).startswith(desc.id):
job.remove()
# add all triggers
if register:
for trigger in desc.triggers:
if isinstance(trigger, EventTrigger):
self.registered_event_trigger.append((trigger, desc))
if isinstance(trigger, TimeTrigger):
uid = f"{desc.id}_{trigger.cron_expression}"
name = f"Trigger for task {desc.id} on cron expression {trigger.cron_expression}"
self.scheduler.cron(uid, name, trigger.cron_expression, self.time_triggered, desc, trigger)
# recompute the lookup table
self.registered_event_trigger_by_message_type = group_by(
lambda t: t[0].message_type, self.registered_event_trigger
)
# task descriptors can hold placeholders (e.g. @NOW@)
# which should be replaced, when the task is started (or restarted).
def evaluate_task_definition(self, descriptor: TaskDescription, **env: str) -> TaskDescription:
def evaluate(step: Step) -> Step:
if isinstance(step.action, ExecuteCommand):
update = copy(step)
update.action = ExecuteCommand(self.cli.replace_placeholder(step.action.command, **env))
return update
else:
return step
updated = copy(descriptor)
updated.steps = [evaluate(step) for step in descriptor.steps]
return updated
async def start_task(self, desc: TaskDescription) -> Optional[RunningTask]:
existing = first(lambda x: x.descriptor.id == desc.id and x.is_active, self.tasks.values()) # type: ignore
if existing:
if desc.on_surpass == TaskSurpassBehaviour.Skip:
log.info(
f"Task {desc.name} has been triggered. Since the last job is not finished, "
f"the execution will be skipped, as defined by the task"
)
return None
elif desc.on_surpass == TaskSurpassBehaviour.Replace:
log.info(f"New task {desc.name} should replace existing run: {existing.id}.")
existing.end()
await self.store_running_task_state(existing)
elif desc.on_surpass == TaskSurpassBehaviour.Parallel:
log.info(f"New task {desc.name} will race with existing run {existing.id}.")
else:
raise AttributeError(f"Surpass behaviour not handled: {desc.on_surpass}")
updated = self.evaluate_task_definition(desc)
wi, commands = RunningTask.empty(updated, self.subscription_handler.subscribers_by_event)
log.info(f"Start new task: {updated.name} with id {wi.id}")
# store initial state in database
await self.running_task_db.insert(wi)
self.tasks[wi.id] = wi
await self.execute_task_commands(wi, commands)
return wi
async def start_interrupted_tasks(self) -> List[RunningTask]:
descriptions = {w.id: w for w in self.task_descriptions}
def reset_state(wi: RunningTask, task_data: RunningTaskData) -> RunningTask:
# reset the received messages
wi.received_messages = task_data.received_messages # type: ignore
# move the fsm into the last known state
wi.machine.set_state(task_data.current_state_name)
# import state of the current step
wi.current_state.import_state(task_data.current_state_snapshot)
# reset times
wi.task_started_at = task_data.task_started_at
wi.step_started_at = task_data.step_started_at
# ignore all messages that would be emitted
wi.move_to_next_state()
return wi
instances: List[RunningTask] = []
async for data in self.running_task_db.all():
descriptor = descriptions.get(data.task_descriptor_id)
if descriptor:
# we have captured the timestamp when the task has been started
updated = self.evaluate_task_definition(descriptor, now=utc_str(data.task_started_at))
rt = RunningTask(data.id, updated, self.subscription_handler.subscribers_by_event)
instance = reset_state(rt, data)
if isinstance(instance.current_step.action, RestartAgainStepAction):
log.info(f"Restart interrupted action: {instance.current_step.action}")
await self.execute_task_commands(instance, instance.current_state.commands_to_execute())
instances.append(instance)
else:
log.warning(f"No task description with this id found: {data.task_descriptor_id}. Remove instance data.")
await self.running_task_db.delete(data.id)
return instances
async def __aenter__(self) -> TaskHandler:
log.info("TaskHandler is starting up!")
# load job descriptions from configuration files
file_jobs = [await self.parse_job_file(file) for file in self.args.jobs] if self.args.jobs else []
jobs: List[Job] = reduce(lambda r, l: r + l, file_jobs, [])
# load job descriptions from database
db_jobs = [job async for job in self.job_db.all()]
self.task_descriptions = [*self.task_descriptions, *jobs, *db_jobs]
# load and restore all tasks
self.tasks = {wi.id: wi for wi in await self.start_interrupted_tasks()}
await self.timeout_watcher.start()
for descriptor in self.task_descriptions:
await self.update_trigger(descriptor)
if self.args.start_collect_on_subscriber_connect:
filtered = [wf for wf in self.known_workflows() if wf.id == "collect"]
self.initial_start_workflow_task = wait_and_start(filtered, self, self.message_bus)
async def listen_to_message_bus() -> None:
async with self.message_bus.subscribe("task_handler") as messages:
while True:
message = None
try:
message = await messages.get()
if isinstance(message, Event):
await self.handle_event(message)
elif isinstance(message, Action):
await self.handle_action(message)
elif isinstance(message, (ActionDone, ActionError)):
log.info(f"Ignore message via event bus: {message}")
except asyncio.CancelledError as ex:
# if we outer task is cancelled, give up
raise ex
except Exception as ex:
log.error(f"Could not handle event {message} - give up.", exc_info=ex)
self.message_bus_watcher = asyncio.create_task(listen_to_message_bus())
return self
async def __aexit__(self, exc_type: Any, exc_val: Any, exc_tb: Any) -> None:
log.info("Tear down task handler")
# deregister from all triggers
for descriptor in self.task_descriptions:
await self.update_trigger(descriptor, register=False)
# stop timeout watcher
await self.timeout_watcher.stop()
# stop event listener
if self.message_bus_watcher:
self.message_bus_watcher.cancel()
try:
await self.message_bus_watcher
except CancelledError:
log.info("task has been cancelled")
# wait for all running commands to complete
for task in list(self.tasks.values()):
if task.update_task:
await task.update_task
del self.tasks[task.id]
# in case the task is not done
if self.initial_start_workflow_task and not self.initial_start_workflow_task.done():
self.initial_start_workflow_task.cancel()
# endregion
# region job handler
async def running_tasks(self) -> List[RunningTask]:
return list(self.tasks.values())
async def start_task_by_descriptor_id(self, uid: str) -> Optional[RunningTask]:
td = first(lambda t: t.id == uid, self.task_descriptions) # type: ignore # pypy
if td:
return await self.start_task(td)
else:
raise NameError(f"No task with such id: {uid}")
async def list_jobs(self) -> List[Job]:
return [job for job in self.task_descriptions if isinstance(job, Job)]
async def add_job(self, job: Job) -> None:
descriptions = list(self.task_descriptions)
existing = first(lambda td: td.id == job.id, descriptions) # type: ignore # pypy
if existing:
if not existing.mutable:
raise AttributeError(f"There is an existing job with this {job.id} which can not be deleted!")
log.info(f"Job with id {job.id} already exists. Update this job.")
descriptions.remove(existing)
# store in database
await self.job_db.update(job)
descriptions.append(job)
self.task_descriptions = descriptions
await self.update_trigger(job)
async def delete_running_task(self, task: RunningTask) -> None:
task.descriptor_alive = False
# remove tasks from list of running tasks
self.tasks.pop(task.id, None)
if task.update_task and not task.update_task.done():
task.update_task.cancel()
# mark step as error
task.end()
# remove from database
await self.running_task_db.delete(task.id)
async def delete_job(self, job_id: str) -> Optional[Job]:
job: Job = first(lambda td: td.id == job_id and isinstance(td, Job), self.task_descriptions) # type: ignore
if job:
if not job.mutable:
raise AttributeError(f"Can not delete job: {job.id} - it is defined in a system file!")
# delete all running tasks of this job
for task in list(filter(lambda x: x.descriptor.id == job.id, self.tasks.values())):
log.info(f"Job: {job_id}: delete running task: {task.id}")
await self.delete_running_task(task)
await self.job_db.delete(job_id)
descriptions = list(self.task_descriptions)
descriptions.remove(job)
self.task_descriptions = descriptions
await self.update_trigger(job, register=False)
return job
# endregion
# region maintain running tasks
async def time_triggered(self, descriptor: TaskDescription, trigger: TimeTrigger) -> None:
log.info(f"Task {descriptor.name} triggered by time: {trigger.cron_expression}")
await self.start_task(descriptor)
async def check_for_task_to_start_on_message(self, msg: Message) -> None:
# check if this event triggers any new task
for trigger, descriptor in self.registered_event_trigger_by_message_type.get(msg.message_type, []):
if msg.message_type == trigger.message_type:
comp = trigger.filter_data
if {key: msg.data.get(key) for key in comp} == comp if comp else True:
log.info(f"Event {msg.message_type} triggers task: {descriptor.name}")
await self.start_task(descriptor)
async def handle_event(self, event: Event) -> None:
# check if any running task want's to handle this event
for wi in list(self.tasks.values()):
handled, commands = wi.handle_event(event)
if handled:
await self.execute_task_commands(wi, commands, event)
# check if this event triggers any new task
await self.check_for_task_to_start_on_message(event)
# noinspection PyMethodMayBeStatic
async def handle_action(self, action: Action) -> None:
await self.check_for_task_to_start_on_message(action)
async def handle_action_result(
self, done: Union[ActionDone, ActionError], fn: Callable[[RunningTask], Sequence[TaskCommand]]
) -> None:
wi = self.tasks.get(done.task_id)
if wi:
commands = fn(wi)
return await self.execute_task_commands(wi, commands, done)
else:
log.warning(
f"Received an ack for an unknown task={done.task_id} "
f"event={done.message_type} from={done.subscriber_id}. Ignore."
)
async def handle_action_done(self, done: ActionDone) -> None:
return await self.handle_action_result(done, lambda wi: wi.handle_done(done))
async def handle_action_error(self, err: ActionError) -> None:
log.info(f"Received error: {err.error} {err.step_name}:{err.message_type} of {err.task_id}")
return await self.handle_action_result(err, lambda wi: wi.handle_error(err))
async def execute_task_commands(
self, wi: RunningTask, commands: Sequence[TaskCommand], origin_message: Optional[Message] = None
) -> None:
async def execute_commands() -> None:
# execute and collect all task commands
results: Dict[TaskCommand, Any] = {}
for command in commands:
if isinstance(command, SendMessage):
await self.message_bus.emit(command.message)
results[command] = None
elif isinstance(command, ExecuteOnCLI):
# TODO: instead of executing it in process, we should do an http call here to a worker core.
result = await self.cli.execute_cli_command(command.command, stream.list, **command.env)
results[command] = result
else:
raise AttributeError(f"Does not understand this command: {wi.descriptor.name}: {command}")
# The descriptor might be removed in the mean time. If this is the case stop execution.
if wi.descriptor_alive:
active_before_result = wi.is_active
# before we move on, we need to store the current state of the task (or delete if it is done)
await self.store_running_task_state(wi, origin_message)
# inform the task about the result, which might trigger new tasks to execute
new_commands = wi.handle_command_results(results)
if new_commands:
# note: recursion depth is defined by the number of steps in a job description and should be safe.
await self.execute_task_commands(wi, new_commands)
elif active_before_result and not wi.is_active:
# if this was the last result the task was waiting for, delete the task
await self.store_running_task_state(wi, origin_message)
async def execute_in_order(task: Task) -> None: # type: ignore # pypy
# make sure the last execution is finished, before the new execution starts
await task
await execute_commands()
# start execution of commands in own task to not block the task handler
# note: the task is awaited finally in the timeout handler or context handler shutdown
wi.update_task = asyncio.create_task(execute_in_order(wi.update_task) if wi.update_task else execute_commands())
async def store_running_task_state(self, wi: RunningTask, origin_message: Optional[Message] = None) -> None:
if wi.is_active:
await self.running_task_db.update_state(wi, origin_message)
elif wi.id in self.tasks:
await self.running_task_db.delete(wi.id)
async def list_all_pending_actions_for(self, subscriber: Subscriber) -> List[Action]:
pending = map(lambda x: x.pending_action_for(subscriber), self.tasks.values())
return [x for x in pending if x]
# endregion
# region periodic task checker
async def check_overdue_tasks(self) -> None:
"""
Called periodically by the system.
In case there is an overdue task, an action error is injected into the task.
"""
for task in list(self.tasks.values()):
if task.is_active: # task is still active
if task.current_state.check_timeout():
if task.current_step.on_error == StepErrorBehaviour.Continue:
current_step = task.current_step.name
commands = task.move_to_next_state()
log.warning(
f"Task {task.id}: {task.descriptor.name} timed out in step "
f"{current_step}. Moving on to step: {task.current_step.name}."
)
await self.execute_task_commands(task, commands)
else:
log.warning(
f"Task {task.id}: {task.descriptor.name} timed out "
f"in step {task.current_step.name}. Stop the task."
)
task.end()
await self.store_running_task_state(task)
# check again for active (might have changed for overdue tasks)
if not task.is_active:
if task.update_task:
await task.update_task
del self.tasks[task.id]
await self.running_task_db.delete(task.id)
# endregion
# region parse job data
async def parse_job_file(self, file: TextIOWrapper) -> List[Job]:
"""
Parse a file with job definitions.
Every line is either a blank line, a comment or a job definition.
Example
# cron based trigger
0 5 * * sat : reported name="foo" | desire name="bla"
# cron based + event based trigger
0 5 * * sat event_name : reported name="foo" | desire name="bla"
:param file: the file handle to parse.
:return: all parsed jobs.
:raises: ParseError if the job can not be parsed
"""
jobs = []
with file as reader:
for line in reader:
stripped = line.strip()
if stripped and not stripped.startswith("#"):
job = await self.parse_job_line(f"file {file.name}", stripped, mutable=False)
jobs.append(job)
return jobs
async def parse_job_line(self, source: str, line: str, mutable: bool = True) -> Job:
"""
Parse a single job line.
:param source: the source of this line (just for naming purposes)
:param line: the line of text
:param mutable: defines if the resulting job is mutable or not.
:return: the parsed jon
"""
stripped = line.strip()
uid = uuid_str(stripped)[0:8]
timeout = timedelta(hours=1)
wait_timeout = timedelta(hours=24)
async def parse_with_cron() -> Job:
parts = re.split("\\s+", stripped, 5)
if len(parts) != 6:
raise ValueError(f"Invalid job {stripped}")
wait: Optional[Tuple[EventTrigger, timedelta]] = None
trigger = TimeTrigger(" ".join(parts[0:5]))
command = parts[5]
# check if we also need to wait for an event: name_of_event : command
if self.event_re.match(command):
event, command = re.split("\\s*:\\s*", command, 1)
wait = EventTrigger(event), wait_timeout
await self.cli.evaluate_cli_command(command, replace_place_holder=False)
return Job(uid, ExecuteCommand(command), trigger, timeout, wait, mutable)
async def parse_event() -> Job:
event, command = re.split("\\s*:\\s*", stripped, 1)
await self.cli.evaluate_cli_command(command, replace_place_holder=False)
return Job(uid, ExecuteCommand(command), EventTrigger(event), timeout, mutable=mutable)
try:
return await (parse_event() if self.event_re.match(stripped) else parse_with_cron())
except CLIParseError as ex:
raise ex
except Exception as ex:
raise ParseError(f"Can not parse job command line: {stripped}") from ex
event_re = re.compile("^[A-Za-z][A-Za-z0-9_\\-]*\\s*:")
# endregion
# region known task descriptors
@staticmethod
def known_jobs() -> List[Job]:
return [
Job(
"example-job",
ExecuteCommand("echo hello"),
EventTrigger("run_job"),
timedelta(seconds=10),
mutable=False,
),
Job(
"example-wait-job",
ExecuteCommand("sleep 10; echo I was started at @NOW@"),
EventTrigger("run_job"),
timedelta(seconds=10),
(EventTrigger("wait"), timedelta(seconds=30)),
mutable=False,
),
]
@staticmethod
def known_workflows() -> List[Workflow]:
collect_steps = [
Step("pre_collect", PerformAction("pre_collect"), timedelta(seconds=10)),
Step("collect", PerformAction("collect"), timedelta(seconds=10)),
Step("post_collect", PerformAction("post_collect"), timedelta(seconds=10)),
]
cleanup_steps = [
Step("pre_plan", PerformAction("pre_cleanup_plan"), timedelta(seconds=10)),
Step("plan", PerformAction("cleanup_plan"), timedelta(seconds=10)),
Step("post_plan", PerformAction("post_cleanup_plan"), timedelta(seconds=10)),
Step("pre_clean", PerformAction("pre_cleanup"), timedelta(seconds=10)),
Step("clean", PerformAction("cleanup"), timedelta(seconds=10)),
Step("post_clean", PerformAction("post_cleanup"), timedelta(seconds=10)),
]
metrics_steps = [
Step("pre_metrics", PerformAction("pre_generate_metrics"), timedelta(seconds=10)),
Step("metrics", PerformAction("generate_metrics"), timedelta(seconds=10)),
Step("post_metrics", PerformAction("post_generate_metrics"), timedelta(seconds=10)),
]
return [
Workflow(
uid="collect",
name="collect",
steps=collect_steps + metrics_steps,
triggers=[EventTrigger("start_collect_workflow")],
),
Workflow(
uid="cleanup",
name="cleanup",
steps=cleanup_steps + metrics_steps,
triggers=[EventTrigger("start_cleanup_workflow")],
),
Workflow(
uid="metrics",
name="metrics",
steps=metrics_steps,
triggers=[EventTrigger("start_metrics_workflow")],
),
Workflow(
uid="collect_and_cleanup",
name="collect_and_cleanup",
steps=collect_steps + cleanup_steps + metrics_steps,
triggers=[EventTrigger("start_collect_and_cleanup_workflow"), TimeTrigger("0 * * * *")],
),
]
# endregion
| [
"noreply@github.com"
] | noreply@github.com |
ffb81ac3e424652e5a48fdb526bf4d905bbd1018 | cfb61b4a2fbe0f7a162e71208b8b772e4300b66b | /read_record/utils.py | 2dae7fcfe0490756c835e00583026a19f3b23923 | [] | no_license | zy421612918/blog | 4c18ca6364fc1d4425dbba82fedb42d762e93ffe | f5b92452b468a54108c23a4a45b6e8c504a3f301 | refs/heads/master | 2020-03-22T12:38:56.977149 | 2018-07-07T04:35:41 | 2018-07-07T04:35:41 | 140,053,094 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,329 | py | #! /usr/bin/env python
# -*- coding: utf-8 -*-
# __author__ = "Zhangye"
# Date: 18-6-6
import datetime
from django.contrib.contenttypes.models import ContentType
from .models import ReadNum,ReadDetail
from django.utils import timezone
from django.db.models import Sum
from Blog.models import Blog
def check_read_record(request,obj):
"""
检查是否阅读过
:param request:
:param obj: blog对象
:return: cookie标记
"""
ct = ContentType.objects.get_for_model(obj)
key = '%s_%s_read'%(ct.model, obj.pk) # blog_35_read
if not request.COOKIES.get(key):
# 总阅读数+1
readnum, created = ReadNum.objects.get_or_create(content_type=ct, object_id=obj.pk)
readnum.read_num += 1
readnum.save()
# 当日计数+1
date = timezone.now().date()
readDetail,created = ReadDetail.objects.get_or_create(content_type=ct, object_id=obj.pk, date=date)
readDetail.read_num+=1
readDetail.save()
return key
def get_sevendays_read(content_type):
today = timezone.now().date()
read_list=[]
date_list=[]
print(today)
for i in range(7,0,-1):
date = today - datetime.timedelta(days=i)
date_list.append(date.strftime('%m/%d'))
read_details=ReadDetail.objects.filter(content_type=content_type, date=date)
res = read_details.aggregate(read_nums=Sum('read_num'))
read_list.append(res['read_nums'] or 0)
return read_list,date_list
def get_hot_data(content_type):
today = timezone.now().date()
read_details = ReadDetail.objects.filter(content_type=content_type, date=today).order_by('-read_num')
return read_details[:3]
def yesterday_hot_data(content_type):
today = timezone.now().date()
yesterday = today- datetime.timedelta(days=1)
read_details = ReadDetail.objects.filter(content_type=content_type, date=yesterday).order_by('-read_num')
return read_details[:3]
def get_7days_hot_date():
today = timezone.now().date()
date = today - datetime.timedelta(days=7)
blogs = Blog.objects.filter(read_details__date__lt=today,
read_details__date__gte=date).\
values('id','title')\
.annotate(read_num_sum=Sum('read_details__read_num')).order_by('-read_num_sum')
return blogs[:7]
| [
"421612918"
] | 421612918 |
257b7003bc935a62b341351ce28e8be2d272b6ae | bb88a96f31e12007ec99f8e64045e790831805fe | /gringotts/api/v1/sub.py | 71f38bf767d8eab916dcde5a1fd699fbbcb865f1 | [] | no_license | rogeroger-yu/ustack-gringotts | 38ed680f111c262a666f55609074a00fd6f76b7e | 75f656398c11b0dbddf99bf429994624915c3565 | refs/heads/master | 2022-09-24T03:52:46.452502 | 2017-03-30T11:03:54 | 2017-03-30T11:19:22 | 86,901,913 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,502 | py | import calendar
import pecan
import wsme
import datetime
import decimal
from pecan import rest
from pecan import request
from wsmeext.pecan import wsexpose
from wsme import types as wtypes
from oslo_config import cfg
from gringotts import exception
from gringotts import utils as gringutils
from gringotts.api.v1 import models
from gringotts.db import models as db_models
from gringotts.openstack.common import log
from gringotts.openstack.common import timeutils
from gringotts.openstack.common import uuidutils
LOG = log.getLogger(__name__)
class SubsController(rest.RestController):
"""The controller of resources
"""
@wsexpose(models.Subscription, body=models.SubscriptionPostBody)
def post(self, data):
conn = pecan.request.db_conn
subscription = conn.create_subscription(request.context,
**data.as_dict())
if subscription:
return models.Subscription.from_db_model(subscription)
else:
return None
@wsexpose(None, body=models.SubscriptionPutBody)
def put(self, data):
conn = pecan.request.db_conn
if data.quantity != wtypes.Unset:
conn.update_subscription(request.context,
**data.as_dict())
elif data.new_flavor != wtypes.Unset and data.old_flavor != wtypes.Unset:
conn.update_flavor_subscription(request.context,
**data.as_dict())
| [
"guangyu@unitedstack.com"
] | guangyu@unitedstack.com |
21be216193f7f5ea64fdfa525d451bf6e7d4d223 | 202c1ceb861171aeae9a63a23ab0f823773a02c5 | /tiny/scheduling/scheduler.py | 61d5bd7e2012769b2316a6066f189b23f53ad0db | [] | no_license | oampo/tiny | 33d62eb1ed6fdf5d015f7a75c3ed07ecb6c7e387 | 83f987e49152bc6716a7cc80c9d3e5ac80e2ce34 | refs/heads/master | 2016-08-03T09:17:22.544009 | 2015-03-13T15:27:37 | 2015-03-13T15:27:37 | 31,021,320 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,423 | py | import time
import asyncio
import functools
from heapq import heappush, heappop
from .patterns import p_dict
class Scheduler:
duration = "duration"
def __init__(self, bpm=120):
self._queue = []
self.bpm = bpm
self._time = time.time()
self._beat = 0
self._beat_in_bar = 0
self._bar = 0
self.beats_per_bar = 4
self._last_beat_time = self._time
self._beat_length = 60 / bpm
self._future = None
def add(self, time, callback, **patterns):
self._add(time, {"callback": callback, "patterns": p_dict(patterns)()})
def _add(self, time, data):
event = (time, data)
heappush(self._queue, event)
if self._queue[0] is event:
self._set_timer()
def add_relative(self, beats, callback, **patterns):
self._update_time()
self.add(self._time + beats * self._beat_length, callback, **patterns)
def add_absolute(self, beat, callback, **patterns):
self._update_time()
time = self._last_beat_time + (beat - self._beat) * self._beat_length
if time < self._time:
return
self.add(time, callback, **patterns)
def play(self, callback, **patterns):
process_dict = functools.partial(self._process_dict,
callback=callback)
self.add(self.time, process_dict, **patterns)
@property
def time(self):
self._update_time()
return self._time
@property
def beat(self):
self._update_time()
return self._beat
@property
def bar(self):
self._update_time()
return self._bar
@property
def beat_in_bar(self):
self._update_time()
return self._beat_in_bar
def _update_time(self):
self._time = time.time()
beats_elapsed = ((self._time - self._last_beat_time) //
self._beat_length)
bars_elapsed = beats_elapsed // self.beats_per_bar
self._beat += beats_elapsed
self._beat_in_bar = ((self._beat_in_bar + beats_elapsed) %
self.beats_per_bar)
self._bar += bars_elapsed
self._last_beat_time += beats_elapsed * self._beat_length
def _process_events(self):
self._update_time()
while len(self._queue) and self._queue[0][0] <= self._time:
event = heappop(self._queue)
event[1]["time"] = event[0]
self._process_event(event[1])
self._set_timer()
def _set_timer(self):
if self._future is not None:
self._future.cancel()
if not len(self._queue):
return
self._future = asyncio.async(self._run(self._queue[0][0]))
@asyncio.coroutine
def _run(self, time):
# Uses wait strategy from threading wait
delay = 0.0005
while self.time < time:
delay = min(delay * 2, time - self._time, 0.05)
yield from asyncio.sleep(delay)
self._process_events()
def _process_event(self, event):
patterns = event["patterns"]
try:
values = next(patterns)
except StopIteration:
return
duration = event["callback"](**values)
if duration:
time = event.pop("time")
self._add(time + duration * self._beat_length, event)
def _process_dict(self, callback, **values):
duration = values.pop(self.duration)
if not callback(**values):
return duration
# Should maybe be part of expression/unit
def _set_parameters(self, patterns):
try:
values = next(patterns)
except StopIteration:
return True
for parameter, value in values.items():
value >> parameter
def _tick_in_dict(self, dict):
duration = dict.pop(self.duration)
dict = {
self.duration: duration,
"patterns": p_dict(dict)()
}
self.play(self._set_parameters, **dict)
def __rrshift__(self, other):
if hasattr(other, "__getitem__"):
return self._tick_in_dict(other)
return NotImplemented
if __name__ == "__main__":
from .. import main
from .patterns import p_iterable
scheduler = Scheduler()
scheduler.play(lambda x: print(x), x=p_iterable([1, 2, 3]) * 2, duration=1)
main()
| [
"joe@oampo.co.uk"
] | joe@oampo.co.uk |
a33ea344425501fccf20a8502fc44380fce73c76 | ebfcae1c5ba2997b2ac4471d5bedc3f5daffcb31 | /dino-master/dino/rest/resources/send.py | a63978198fcff73e1c60cefb3ad6386d3ea9a807 | [
"Apache-2.0"
] | permissive | babiato/flaskapp1 | 84de2d0b26a54f5820d3bbe97926782ad41e005c | 530beb9e3b8516e0e93960b99521c23a523ef546 | refs/heads/master | 2023-02-26T16:36:49.760632 | 2021-02-04T09:08:40 | 2021-02-04T09:08:40 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,723 | py | import logging
import traceback
import eventlet
import sys
from dino import environ
from dino import utils
from dino.utils.decorators import timeit
from dino.db.manager import UserManager
from dino.rest.resources.base import BaseResource
from flask import request
logger = logging.getLogger(__name__)
def fail(error_message):
return {
'status': 'FAIL',
'message': error_message
}
class SendResource(BaseResource):
def __init__(self):
super(SendResource, self).__init__()
self.user_manager = UserManager(environ.env)
self.request = request
def async_post(self, json):
logger.debug('POST request: %s' % str(json))
if 'content' not in json:
raise RuntimeError('no key [content] in json message')
msg_content = json.get('content')
if msg_content is None or len(msg_content.strip()) == 0:
raise RuntimeError('content may not be blank')
if not utils.is_base64(msg_content):
raise RuntimeError('content in json message must be base64')
user_id = str(json.get('user_id', 0))
user_name = utils.b64d(json.get('user_name', utils.b64e('admin')))
object_type = json.get('object_type')
target_id = str(json.get('target_id'))
namespace = json.get('namespace', '/ws')
target_name = json.get('target_name')
data = utils.activity_for_message(user_id, user_name)
data['target'] = {
'objectType': object_type,
'id': target_id,
'displayName': target_name,
'url': namespace
}
data['object'] = {
'content': msg_content
}
if not environ.env.cache.user_is_in_multicast(target_id):
logger.info('user {} is offline, dropping message: {}'.format(target_id, str(json)))
return
try:
environ.env.out_of_scope_emit('message', data, room=target_id, json=True, namespace='/ws', broadcast=True)
except Exception as e:
logger.error('could not /send message to target {}: {}'.format(target_id, str(e)))
logger.exception(traceback.format_exc())
environ.env.capture_exception(sys.exc_info())
@timeit(logger, 'on_rest_send')
def do_post(self):
is_valid, msg, json = self.validate_json(self.request, silent=False)
if not is_valid:
logger.error('invalid json: %s' % msg)
raise RuntimeError('invalid json')
if json is None:
raise RuntimeError('no json in request')
if not isinstance(json, dict):
raise RuntimeError('need a dict')
eventlet.spawn_n(self.async_post, dict(json))
| [
"jinxufang@tencent.com"
] | jinxufang@tencent.com |
fc5950f4de4608412e324da17fb6c26e9e26f0c3 | 637175e6a58817d77ce160744de1ee1b43d76cda | /tutorial/tutorial/spiders/TorDown.py | a3abf30b597692d36831756b899d4a9b6b04c9bf | [] | no_license | AgentG2015/Scrapy | 283a26f7201deaa1498a429fda77421f470f93f1 | 30687f75c578ccb5d5c0c488d639567ce8013406 | refs/heads/master | 2020-04-06T04:08:47.260255 | 2016-07-04T06:33:51 | 2016-07-04T06:33:51 | 59,291,648 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,006 | py | import scrapy
from tutorial.items import *
from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.http import Request,FormRequest
from scrapy.utils.project import get_project_settings
class TorDown(CrawlSpider):
name = "tordown"
allowed_domains = ["tp.m-team.cc"]
link_template = r"https://tp.m-team.cc/download.php?id="
items = []
start_urls = []
def __init__(self):
self.headers = HEADER
self.cookies = COOKIES
with open('items.txt', 'r') as f:
self.items = f.read().split("\n")[0:-1]
print len(self.items), "items"
def start_requests(self):
for i, id in enumerate(self.items):
#if i > 9: break
url = self.link_template + id
request = FormRequest(url,
headers = self.headers,
cookies =self.cookies,
callback = self.parse_item)
request.meta['id'] = id
yield request
def parse_item(self, response):
filename = "torrents/" + response.meta['id'] + ".torrent"
with open(filename, 'wb') as f:
f.write(response.body)
HEADER={
"Host": "tp.m-team.cc",
"Connection": "keep-alive",
"Cache-Control": "max-age=0",
"Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8",
"User-Agent": "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/40.0.2214.115 Safari/537.36",
"Referer": "https://tp.m-team.cc/torrents.php?sort=5&type=desc",
"Accept-Encoding": "gzip, deflate, sdch",
"Accept-Language": "zh-CN,zh;q=0.8,en;q=0.6,ja;q=0.4,zh-TW;q=0.2"
}
COOKIES={
'c_lang_folder':r'en',
'c_secure_uid':r'MTU5NTI2',
'c_secure_pass':r'd310dc816dd70e1425f042d8a6dc5b1c',
'c_secure_ssl':r'eWVhaA%3D%3D',
'c_secure_tracker_ssl':r'eWVhaA%3D%3D',
'c_secure_login':r'bm9wZQ%3D%3D',
}
| [
"AgentG2015@gmail.com"
] | AgentG2015@gmail.com |
cc43440bf8eef90f605cad03be1c751c39264596 | dbc5bfb2b3eec582c58dc3e76c1e92450250b371 | /app.py | a39b092802bf44862baf5865cdc4c07b315988d4 | [] | no_license | olawrence799/webScraping | ac36a0bfa5ce15fe506e9ebbfbfcebc625014e17 | 50e608426ebfe1efbc237de16ba30c21f5b19553 | refs/heads/master | 2021-09-09T16:38:54.013295 | 2018-03-18T05:03:49 | 2018-03-18T05:03:49 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 511 | py | from flask import Flask, render_template, jsonify, redirect
from flask_pymongo import PyMongo
import scrape_mars
app = Flask(__name__)
mongo = PyMongo(app)
@app.route('/')
def index():
mars = mongo.db.mars.find_one()
return render_template('index.html', mars=mars)
@app.route('/scrape')
def scrape():
mars = mongo.db.mars
data = scrape_mars.scrape()
mars.update({}, data, upsert=True)
return redirect("http://localhost:5000/", code=302)
if __name__ == "__main__":
app.run(debug=True) | [
"owen@Owens-MacBook-Pro.local"
] | owen@Owens-MacBook-Pro.local |
aa6a81ca2a68d3bbe0fcd037c5db7068f2adb766 | dd44e145ac547209f5f209bc9b1f09189bb8b5c7 | /Python-Advanced-2021/03.Multidimensional-lists-L/02.Sum-of-matrix-columns.py | 719862b6c9646cc99f509dcebd28edadbfe2e5d6 | [] | no_license | todorovventsi/Software-Engineering | e3c1be8f0f72c85619518bb914d2a4dbaac270f8 | 64ffa6c80b190e7c6f340aaf219986f769f175ab | refs/heads/master | 2023-07-09T05:35:14.522958 | 2021-08-15T14:35:55 | 2021-08-15T14:35:55 | 336,056,643 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 254 | py | rows, columns = map(int, input().split(", "))
matrix = [[int(i) for i in input().split()] for _ in range(rows)]
for column in range(columns):
column_sum = 0
for row in range(rows):
column_sum += matrix[row][column]
print(column_sum)
| [
"todorov.ventsi@gmail.com"
] | todorov.ventsi@gmail.com |
c413c1f3310160c4a5734f569e47321d1e3bd9a6 | 36989b4bcd070b0889e9be921f8142df9cec67f5 | /PYTHON/Questao10.py | 850eed441aa1b2f2d9c2637f0aca160fa34890d1 | [] | no_license | larissacsf/Exercises-1 | 61dfc46d594c3a20673f65809606cdc814639792 | 26f39571ba870c1a584ac8ad80efe841b231018e | refs/heads/master | 2021-05-23T09:48:08.245971 | 2020-04-05T15:46:52 | 2020-04-05T15:46:52 | 253,228,979 | 0 | 0 | null | 2020-04-05T12:25:30 | 2020-04-05T12:25:29 | null | UTF-8 | Python | false | false | 498 | py | '''Elaborar um algoritmo que lê 3 valores a,b,c e os escreve. A seguir, encontre o maior
dos 3 valores e o escreva com a mensagem : "É o maior".'''
def maiorNumero(n1, n2, n3):
if n1 > n2 and n1 > n3:
aux = n1
elif n2 > n3:
aux = n2
else:
aux = n3
return aux
def main():
n1 = int(input("Digite um número: "))
n2 = int(input("Digite um número: "))
n3 = int(input("Digite um número: "))
print(maiorNumero(n1, n2, n3), "é o maior")
main() | [
"larissa_cuane10@hotmail.com"
] | larissa_cuane10@hotmail.com |
d4ef7df593f1fbf7027fa866174ceb80592f6f0c | bb33e6be8316f35decbb2b81badf2b6dcf7df515 | /source/res/scripts/client/gui/battle_control/controllers/quest_progress/__init__.py | f9b0128616646671d06aafd2df3f29f0785e39a0 | [] | no_license | StranikS-Scan/WorldOfTanks-Decompiled | 999c9567de38c32c760ab72c21c00ea7bc20990c | d2fe9c195825ececc728e87a02983908b7ea9199 | refs/heads/1.18 | 2023-08-25T17:39:27.718097 | 2022-09-22T06:49:44 | 2022-09-22T06:49:44 | 148,696,315 | 103 | 39 | null | 2022-09-14T17:50:03 | 2018-09-13T20:49:11 | Python | UTF-8 | Python | false | false | 151 | py | # Python bytecode 2.7 (decompiled from Python 2.7)
# Embedded file name: scripts/client/gui/battle_control/controllers/quest_progress/__init__.py
pass
| [
"StranikS_Scan@mail.ru"
] | StranikS_Scan@mail.ru |
d34c725b1262b78badb8817d31b511b34234cf00 | 9809666dd2cdedf154ed8056fa9795d30307644a | /OPENCV BEGGINNER LEVEL/code/image_processing_2.py | 78711a9ac14e2ceea7733ee2f5b8ef5de5cde5a7 | [] | no_license | draco-git/test | dcdb91eb780926c8a20ac844cb64c6295d14c176 | c0c76c04c265a52a5fb379751adf4fa17eadd646 | refs/heads/master | 2023-08-05T07:10:31.699593 | 2021-09-20T05:59:16 | 2021-09-20T05:59:16 | 408,156,269 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,819 | py | import cv2
import numpy as np
import import_images
import stacked_images
import gui_features
#reading images
images = import_images.dict_images()
names_of_img_in_dict =[ ]
for i,j in images.items():
images[i] = cv2.imread(j)
names_of_img_in_dict.append(i)
#print(names_of_img_in_dict) : ['avengers', 'balls', 'black_ball', 'blue_ball', 'buildings', 'car', 'dragon', 'iphone_logo', 'logo']
#images dict consists of all read images
#geomatric transformations
#.Translation : transformation matrix req is [[1 0 tx],[0,1,ty]]
def translation():
img = images['car']
transformation_matrix = np.float32([[1, 0, 100], [0, 1, 50]])
r, c, ch = img.shape
trans_img = cv2.warpAffine(img, transformation_matrix, (r, c))
cv2.imshow('trans_img', stacked_images.stacked_matrix([[trans_img, img]], 0.9,(512,512),False))
cv2.waitKey(0)
cv2.destroyAllWindows()
#translation()
#rotation: transformation_matrix req is obtained by cv2.getRotationMatrix2D((cols,rows),angle,scale)
def rotaion():
img = images['buildings']
img = cv2.resize(img,(512,512))
M = cv2.getRotationMatrix2D((256,256),45,0.5) # here 0.5 is the scale in which the output image is showed
rot_img = cv2.warpAffine(img,M,(512,512))
cv2.imshow('rotate_img',stacked_images.stacked_matrix([[img,rot_img]],1,(512,512),False))
cv2.waitKey(0)
cv2.destroyAllWindows()
#rotaion()
#Affine transformation : here in the output image the lines are parallel similar to input image
#in this we cv2.getAffineTransform() to give transformation_matrix
def affine_transform():
out = np.ones((512,512,3),np.uint8)
img = images['buildings']
img = cv2.resize(img,(512,512))
r,c,ch = img.shape
points = gui_features.finding_bgr(img)
output_img = cv2.resize(out,(r,c))
out_points = gui_features.finding_bgr(output_img)
print(points,out_points)
M = cv2.getAffineTransform(np.float32([points]),np.float32([out_points]))
print(M)
dst = cv2.warpAffine(img,M,(r,c))
cv2.imshow('ouput',stacked_images.stacked_matrix([[img,dst]],1,(512,512),False))
cv2.waitKey(0)
#affine_transform()
#perspective_transformation : here we use cv2.getPerspectiveTransforms()
def perspective_transform():
out = np.ones((512,512,3),np.uint8)
img = images['cards']
img = cv2.resize(img,(512,512))
points = gui_features.finding_bgr(img)
input_points = np.float32([points])
out_image = cv2.resize(out,(img.shape[0],img.shape[1]))
o_points = gui_features.finding_bgr(out_image)
output_points = np.float32([o_points])
M = cv2.getPerspectiveTransform(input_points,output_points)
dst = cv2.warpPerspective(img,M,(img.shape[0],img.shape[1]))
cv2.imshow('output',stacked_images.stacked_matrix([[img,dst]],0.8,(512,512),False))
cv2.waitKey(0)
cv2.destroyAllWindows()
#perspective_transform()
#blurring images
def nothing(x): # function used for trackbar
pass
def blurring():
img = images['black_ball']
while True:
kernel = int(input('enter kernel value'))
if kernel == 0:
break
guassian_blur = cv2.GaussianBlur(img, (kernel,kernel), 0)
median_blur = cv2.medianBlur(img, kernel) # median_blur makaes the most blurred compared to guassian_blur
bilteral_filter_image = cv2.bilateralFilter(img,11,11,11,cv2.BORDER_DEFAULT) # only used for noise removal and effective
cv2.imshow('images_blurred', stacked_images.stacked_matrix([[img, guassian_blur, median_blur,bilteral_filter_image]], 0.5,(512,512),False))
if cv2.waitKey(0) & 0xFF == ord('s'):
cv2.destroyAllWindows()
#blurring()
# morphological transformations : normally performed on binary images , consist of two methods erosion and dilution.
# erosion:
def morpho_transformations():
img = images['cards']
gray_image = cv2.cvtColor(img,cv2.COLOR_BGR2GRAY)
while True:
kernel = int(input('enter kernel'))
if kernel == 0:
break
#iterations = int(input('enter no of iterations'))
k = np.ones((kernel,kernel),np.uint8)
dilate = cv2.dilate(gray_image,k,iterations)
erode = cv2.erode(gray_image,k,iterations)
opening = cv2.morphologyEx(gray_image,cv2.MORPH_OPEN,k)
closing = cv2.morphologyEx(gray_image, cv2.MORPH_CLOSE, k)
morphology_gradient = cv2.morphologyEx(gray_image,cv2.MORPH_GRADIENT,k)
top_hat = cv2.morphologyEx(gray_image,cv2.MORPH_TOPHAT,k)
black_hat = cv2.morphologyEx(gray_image,cv2.MORPH_BLACKHAT,k)
cv2.imshow('morpho_transformations',stacked_images.stacked_matrix([[opening,closing,morphology_gradient],[top_hat,black_hat]],0.5,(512,512),True))
if cv2.waitKey(0) & 0xFF == ord('s'):
cv2.destroyAllWindows()
#morpho_transformations()
| [
"madhavgedela3@gmail.com"
] | madhavgedela3@gmail.com |
d7e8659136c28db77da24a4ae6108c9816b9874f | 13e406093d44072cf378689ca5c7116b5426701e | /TugasKelas2Pembahasan.py | b1053f43fdbd6ae0b9ef6da914b5d76edcd5f320 | [] | no_license | ReyhanR/fundamental-python | 9251c013493d3567bd4a98da202b14d75a8aa88f | 8042671d6b08ce551a43f4b094a271570f3bbaf7 | refs/heads/master | 2021-01-03T06:26:29.443876 | 2020-02-19T08:38:56 | 2020-02-19T08:38:56 | 239,959,999 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 257 | py | # Ganjil genap
# Genap, habis dibagi 2 atau hasilnya 0 jika di modulus dengan 2 / Ganjil,
number = int(input('Masukkan angka : '))
if (number % 2 == 0):
status = 'Genap'
else:
status = 'Ganjil'
print(f'Angka {number} termasuk bilangan {status}') | [
"rryanafi@gmail.com"
] | rryanafi@gmail.com |
4a8ba21250def0e771eb0d8cfee9b9e5f35ef4b8 | e87415a8507341d66991411c91e77ad38cda3df9 | /templated_email/block_render.py | 9ddf822ec66358a729ae9b8e2ad0a806ddf76d91 | [
"MIT"
] | permissive | somair/django-templated-email | 6185abf24031a9813fc8b9d53faa8433f7bda0a6 | b217a3e38d7af8b514d8f83568c1fd55efd1ac11 | refs/heads/master | 2021-01-19T14:13:10.500289 | 2017-01-13T13:06:06 | 2017-01-13T13:06:06 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,748 | py | from django.template.loader_tags import BlockNode, ExtendsNode
from django.template import loader, Context, RequestContext
def _get_template(template):
if isinstance(template, (tuple, list)):
return loader.select_template(template)
return loader.get_template(template)
class BlockNotFound(Exception):
"""The requested block did not exist."""
pass
def render_template_block(template, block, context):
"""
Renders a single block from a template.
This template should have previously been rendered.
"""
template._render(context)
return _render_template_block_nodelist(template.nodelist, block, context)
def _render_template_block_nodelist(nodelist, block, context):
for node in nodelist:
if isinstance(node, BlockNode) and node.name == block:
return node.render(context)
for key in ('nodelist', 'nodelist_true', 'nodelist_false'):
if hasattr(node, key):
try:
rendered = _render_template_block_nodelist(
getattr(node, key), block, context)
except:
pass
else:
return rendered
for node in nodelist:
if isinstance(node, ExtendsNode):
try:
rendered = render_template_block(
node.get_parent(context), block, context)
except BlockNotFound:
pass
else:
return rendered
raise BlockNotFound
def render_block_to_string(template_name, block, dictionary=None,
context_instance=None):
"""Return a string
Loads the given template_name and renders the given block with the
given dictionary as context.
"""
dictionary = dictionary or {}
t = _get_template(template_name)
if context_instance:
context_instance.update(dictionary)
else:
context_instance = Context(dictionary)
return render_template_block(t, block, context_instance)
def direct_block_to_template(request, template, block, extra_context=None,
mimetype=None, **kwargs):
"""
Render a given block in a given template with any extra URL
parameters in the context as ``{{ params }}``.
"""
if extra_context is None:
extra_context = {}
dictionary = {'params': kwargs}
for key, value in extra_context.items():
if callable(value):
dictionary[key] = value()
else:
dictionary[key] = value
c = RequestContext(request, dictionary)
t = _get_template(template)
t.render(c)
return HttpResponse(render_template_block(t, block, c), mimetype=mimetype)
| [
"alex.hayes@roi.com.au"
] | alex.hayes@roi.com.au |
c797e1ec5b3e5955a867418fed9a26431bd4212c | 6fa701cdaa0d83caa0d3cbffe39b40e54bf3d386 | /google/cloud/redis/v1/redis-v1-py/google/cloud/redis_v1/services/cloud_redis/pagers.py | ea1c2287e22e2c73eb752e030a4919c860621449 | [
"Apache-2.0"
] | permissive | oltoco/googleapis-gen | bf40cfad61b4217aca07068bd4922a86e3bbd2d5 | 00ca50bdde80906d6f62314ef4f7630b8cdb6e15 | refs/heads/master | 2023-07-17T22:11:47.848185 | 2021-08-29T20:39:47 | 2021-08-29T20:39:47 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,709 | py | # -*- coding: utf-8 -*-
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
from typing import Any, AsyncIterable, Awaitable, Callable, Iterable, Sequence, Tuple, Optional
from google.cloud.redis_v1.types import cloud_redis
class ListInstancesPager:
"""A pager for iterating through ``list_instances`` requests.
This class thinly wraps an initial
:class:`google.cloud.redis_v1.types.ListInstancesResponse` object, and
provides an ``__iter__`` method to iterate through its
``instances`` field.
If there are more pages, the ``__iter__`` method will make additional
``ListInstances`` requests and continue to iterate
through the ``instances`` field on the
corresponding responses.
All the usual :class:`google.cloud.redis_v1.types.ListInstancesResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(self,
method: Callable[..., cloud_redis.ListInstancesResponse],
request: cloud_redis.ListInstancesRequest,
response: cloud_redis.ListInstancesResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()):
"""Instantiate the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.redis_v1.types.ListInstancesRequest):
The initial request object.
response (google.cloud.redis_v1.types.ListInstancesResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = cloud_redis.ListInstancesRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
def pages(self) -> Iterable[cloud_redis.ListInstancesResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = self._method(self._request, metadata=self._metadata)
yield self._response
def __iter__(self) -> Iterable[cloud_redis.Instance]:
for page in self.pages:
yield from page.instances
def __repr__(self) -> str:
return '{0}<{1!r}>'.format(self.__class__.__name__, self._response)
class ListInstancesAsyncPager:
"""A pager for iterating through ``list_instances`` requests.
This class thinly wraps an initial
:class:`google.cloud.redis_v1.types.ListInstancesResponse` object, and
provides an ``__aiter__`` method to iterate through its
``instances`` field.
If there are more pages, the ``__aiter__`` method will make additional
``ListInstances`` requests and continue to iterate
through the ``instances`` field on the
corresponding responses.
All the usual :class:`google.cloud.redis_v1.types.ListInstancesResponse`
attributes are available on the pager. If multiple requests are made, only
the most recent response is retained, and thus used for attribute lookup.
"""
def __init__(self,
method: Callable[..., Awaitable[cloud_redis.ListInstancesResponse]],
request: cloud_redis.ListInstancesRequest,
response: cloud_redis.ListInstancesResponse,
*,
metadata: Sequence[Tuple[str, str]] = ()):
"""Instantiates the pager.
Args:
method (Callable): The method that was originally called, and
which instantiated this pager.
request (google.cloud.redis_v1.types.ListInstancesRequest):
The initial request object.
response (google.cloud.redis_v1.types.ListInstancesResponse):
The initial response object.
metadata (Sequence[Tuple[str, str]]): Strings which should be
sent along with the request as metadata.
"""
self._method = method
self._request = cloud_redis.ListInstancesRequest(request)
self._response = response
self._metadata = metadata
def __getattr__(self, name: str) -> Any:
return getattr(self._response, name)
@property
async def pages(self) -> AsyncIterable[cloud_redis.ListInstancesResponse]:
yield self._response
while self._response.next_page_token:
self._request.page_token = self._response.next_page_token
self._response = await self._method(self._request, metadata=self._metadata)
yield self._response
def __aiter__(self) -> AsyncIterable[cloud_redis.Instance]:
async def async_generator():
async for page in self.pages:
for response in page.instances:
yield response
return async_generator()
def __repr__(self) -> str:
return '{0}<{1!r}>'.format(self.__class__.__name__, self._response)
| [
"bazel-bot-development[bot]@users.noreply.github.com"
] | bazel-bot-development[bot]@users.noreply.github.com |
48d347ad45dc5dc9a881be62103d5e32a1581f13 | 27cc8ec982975383a07e7781b2b09600f4a2518d | /demo/urls.py | 54abef98b39afec91417cc5bf6e80175ce307529 | [] | no_license | srilekha201created/Grammer | b038677ee4152261fb1b79a7f03f373fcdfb047e | e8110620a98511f4c6518eb99a262cfc05355985 | refs/heads/master | 2023-04-07T17:43:17.407537 | 2021-04-15T10:14:10 | 2021-04-15T10:14:10 | 358,164,863 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 858 | py | """demo URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/3.1/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: path('', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: path('', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.urls import include, path
2. Add a URL to urlpatterns: path('blog/', include('blog.urls'))
"""
from django.contrib import admin
from django.urls import path
from spellcorrection import views
urlpatterns = [
path('admin/', admin.site.urls),
path('', views.home, name='home'),
path('res',views.res, name='res'),
]
| [
"srilekhajagarlamudi.ce@gmail.com"
] | srilekhajagarlamudi.ce@gmail.com |
cc6ae0e7dab5a85c0321944085e16e0d3c43552e | 10eeee95f5f3fb436a0a3cf50a2fd057779808af | /sockjs/cyclone/__init__.py | 3aa631ff8643b6d96b46cb0628e42c3529c4cb58 | [
"MIT"
] | permissive | flaviogrossi/sockjs-cyclone | b9c84c28036049ee58f948e4cf5e33a850286fb0 | d3ca053ec1aa1e85f652347bff562c2319be37a2 | refs/heads/master | 2020-05-31T07:02:48.428630 | 2014-08-13T07:20:13 | 2014-08-13T07:20:13 | 4,571,140 | 11 | 2 | null | 2014-07-16T19:14:26 | 2012-06-06T10:14:36 | Python | UTF-8 | Python | false | false | 106 | py | from .conn import SockJSConnection
from .router import SockJSRouter
from .conn import MultiplexConnection
| [
"flaviogrossi@gmail.com"
] | flaviogrossi@gmail.com |
d198cc72d424a69a2c2d0a225554d187793a8099 | 2415d7f3cc3a4cb98897e91386d861d4b672e824 | /src/puremvc/utilities/pipes/messages.py | c8189aae63b178e59f8c87c0193b09a84c06a93f | [] | no_license | Yoriz/puremvc-python-util-pipes | cc4cf60675f4e700736a6bbf6bda13374a37d87e | c8136e648683560c991c0ff359eb14e509f84ee2 | refs/heads/master | 2021-04-09T16:54:36.803883 | 2012-11-28T19:16:25 | 2012-11-28T19:16:25 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,646 | py | '''
Created on 26 Nov 2012
@author: Dave Wilson
'''
from puremvc.utilities.pipes import interfaces
class Message(interfaces.IPipeMessage):
"""
Pipe Message.
Messages travelling through a Pipeline can be filtered, and queued.
In a queue, they may be sorted by priority. Based on type, they may used
as control messages to modify the behaviour of filter or queue fittings
connected to the pipeline into which they are written.
"""
PRIORITY_HIGH = 1
"""High priority Messages can be sorted to the front of the queue"""
PRIORITY_MED = 5
"""Medium priority Messages are the default"""
PRIORITY_LOW = 10
"""Low priority Messages can be sorted to the back of the queue"""
BASE = 'http://puremvc.org/namespaces/pipes/messages/'
NORMAL = BASE + 'normal/'
"""Normal Message type."""
def __init__(self, msg_type, header=None, body=None, priority=5):
"""
@param msg_type: Messages can be handled differently according to type
@param header: Header properties describe any meta data about the
message for the recipient
@param body: Body of the message is the precious cargo
@param priority: Messages in a queue can be sorted by priority.
"""
self.msg_type = msg_type
self.header = header
self.body = body
self.priority = priority
def getType(self):
"""Get the type of this message"""
return self.msg_type
def setType(self, msg_type):
"""Set the type of this message"""
self.msg_type = msg_type
def getPriority(self):
"""Get the priority of this message"""
return self.priority
def setPriority(self, priority):
"""Set the priority of this message"""
self.priority = priority
def getHeader(self):
"""Get the header of this message"""
return self.header
def setHeader(self, header):
"""Set the header of this message"""
self.header = header
def getBody(self):
"""Get the body of this message"""
return self.body
def setBody(self, body):
"""Set the body of this message"""
self.body = body
class FilterControlMessage(Message):
"""
Filter Control Message.
A special message type for controlling the behaviour of a Filter.
The FilterControlMessage.SET_PARAMS message type tells the Filter to
retrieve the filter parameters object.
The FilterControlMessage.SET_FILTER message type tells the Filter to
retrieve the filter function.
The FilterControlMessage.BYPASS message type tells the Filter that it
should go into Bypass mode operation, passing all normal messages
through unfiltered.
The FilterControlMessage.FILTER message type tells the Filter that it
should go into Filtering mode operation, filtering all normal normal
messages before writing out. This is the default mode of operation and so
this message type need only be sent to cancel a previous
FilterControlMessage.BYPASS message.
The Filter only acts on a control message if it is targeted to this named
filter instance. Otherwise it writes the message through to its output
unchanged.
"""
BASE = Message.BASE + 'filter-control/'
"""Message type base URI"""
SET_PARAMS = BASE + 'setparams'
"""Set filter parameters."""
SET_FILTER = BASE + 'setfilter'
"""Set filter function."""
BYPASS = BASE + 'bypass'
"""Toggle to filter bypass mode."""
FILTER = BASE + 'filter'
"""Toggle to filtering mode. (default behaviour)."""
def __init__(self, msg_type, name, msg_filter=None, params=None):
"""
@param name: The target filter name.
@param msg_filter: The filter function.
@param params: The parameters object.
"""
super(FilterControlMessage, self).__init__(msg_type)
self.name = name
self.msg_filter = msg_filter
self.params = params
def setName(self, name):
"""Set the target filter name."""
self.name = name
def getName(self):
"""Get the target filter name."""
return self.name
def setFilter(self, msg_filter):
"""Set the filter function."""
self.msg_filter = msg_filter
def getFilter(self):
"""Get the filter function."""
return self.msg_filter
def setParams(self, params):
"""Set the parameters object."""
self.params = params
def getParams(self):
"""Get the parameters object."""
return self.params
class QueueControlMessage(Message):
"""
Queue Control Message.
A special message for controlling the behaviour of a Queue.
When written to a pipeline containing a Queue, the type of the message is
interpreted and acted upon by the Queue.
Unlike filters, multiple serially connected queues aren't very useful and
so they do not require a name. If multiple queues are connected serially,
the message will be acted upon by the first queue only.
"""
BASE = Message.BASE + '/queue/'
FLUSH = BASE + 'flush'
"""Flush the queue."""
SORT = BASE + 'sort'
"""Toggle to sort-by-priority operation mode."""
FIFO = BASE + 'fifo'
"""Toggle to FIFO operation mode (default behaviour)."""
def __init__(self, msg_type):
super(QueueControlMessage, self).__init__(msg_type)
| [
"dave.wilson1@ntlworld.com"
] | dave.wilson1@ntlworld.com |
d47adad75197246557075f605b07f839b2123a11 | d4f1bd5e52fe8d85d3d0263ede936928d5811bff | /Python/Problem Solving/BOJ/boj5724.py | c8ae00d9227f3f31db76e85c3ec8da1ed321f2ee | [] | no_license | ambosing/PlayGround | 37f7d071c4402599995a50cac1e7f1a85c6d10dd | 0d5262dbb2fa2128ecb3fd969244fa647b104928 | refs/heads/master | 2023-04-08T04:53:31.747838 | 2023-03-23T06:32:47 | 2023-03-23T06:32:47 | 143,112,370 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 327 | py | lst = list()
idx_lst = list()
max_val = -1
while True:
n = int(input())
if n == 0:
break
idx_lst.append(n)
if max_val < n:
max_val = n
for i in range(1, max_val + 1):
if i != 1:
lst.append(lst[i - 2] + i ** 2)
else:
lst.append(1)
for i in idx_lst:
print(lst[i - 1])
| [
"jiyang@student.42seoul.kr"
] | jiyang@student.42seoul.kr |
ce6dd7b3572bd458bdf2c0016d8e8241ce7d3b27 | bd4a7496f732d9268377292616f6973586abf7b6 | /login.py | 467bba5e37b9677be8aa671ce35edd98b3b0ea4a | [] | no_license | Code-Institute-Submissions/squirrel | 35853094cc42e0a32a47c4d8a77d12dee09b9c3d | 420022b54af404a3bb1b6580ac52b011d1af8285 | refs/heads/master | 2022-12-09T20:34:13.704136 | 2020-08-24T16:14:30 | 2020-08-24T16:14:30 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 854 | py |
from flask_login import LoginManager, UserMixin
from config import app, users
from bson.objectid import ObjectId
"""
# USER MANAGEMENT
# ===============
# The following uses flask_login to create a user class that will be used
# throughout the application.
"""
login_manager = LoginManager(app)
login_manager.login_view = 'login'
login_manager.login_message_category = 'info'
login_manager.init_app(app)
class User(UserMixin):
def __init__(self, user):
self.user = user
self.username = user['username']
self.id = user['_id']
self.email = user['email']
self.password = user['password']
def get_id(self):
object_id = self.user['_id']
return str(object_id)
@login_manager.user_loader
def load_user(user_id):
user = users.find_one({'_id': ObjectId(user_id)})
return User(user)
| [
"simoncastagna@Simons-MacBook-Pro.local"
] | simoncastagna@Simons-MacBook-Pro.local |
6c685c912dc3606c91cb0c2ed32d5fd4d5107297 | 0e2005613bfbefaa5dd423f74326cf7854f7d83c | /UbuntuScript2.0/script.py | 6d5ff2d0f2347ed7f61e7ec65e856b663e0dc466 | [] | no_license | xrchsploit/Tricks | 1f3bcf4e3f73ece23dba6ef495bdc370ae7b86e3 | 8a7753bf0c5b881cc09f4a266dee06395e37da69 | refs/heads/main | 2023-02-20T05:48:24.911850 | 2021-01-23T03:20:27 | 2021-01-23T03:20:27 | 325,204,038 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 8,149 | py | # Modules that are going to be used
import os
import time
# Colors for text
class bcolors:
RED = '\033[91m'
BLUE = '\033[94m'
GREEN = '\033[92m'
YELLOW = '\033[93m'
NORMAL = '\033[0m'
BOLD = '\033[1m'
FLASH = '\033[5m'
os.system("clear")
# Firefox settings, warning, apt fix, GUI updates,
print(bcolors.RED + bcolors.FLASH + "Ensure you ran this as root. I.E (sudo python3 script.py) " + bcolors.NORMAL)
print(bcolors.GREEN + "Ensure you did things like: fix apt and got the GUI software update settings " + bcolors.NORMAL)
print('')
print(bcolors.GREEN + "Ensure you go to firefox > privacy and security settings > enable delete cookies > disable ask to save logins" + bcolors.NORMAL)
print(bcolors.GREEN + "enable block pop ups > Block deceptive content" + bcolors.NORMAL)
print("")
print(bcolors.RED + bcolors.FLASH + "Screenshot all logins! this will do PAM and LightDM!!!" + bcolors.NORMAL)
time.sleep(15)
# Fixing package manager, any broken dependencies will be fixed
os.system("sudo dpkg --configure -a")
os.system("sudo apt clean")
# Updates and installing nessesary tools
os.system("sudo apt install net-tools -y")
os.system("sudo apt update -y && sudo apt upgrade -y")
os.system("sudo apt install clamav -y")
os.system("sudo apt install synaptic -y")
os.system("sudo apt install ufw -y")
os.system("sudo apt install gufw -y")
os.system("sudo apt install rkhunter -y")
os.system("sudo apt install gedit -y")
os.system("sudo apt install terminator -y")
# Bum may not work, hit or miss
os.system("sudo apt install bum")
os.system("clear")
os.system("sudo apt install libpam-cracklib -y")
cracklib = input(bcolors.YELLOW + "*Was cracklib installed correctly? If no tyne 'n' and go fix it! y/n * " + bcolors.NORMAL)
# Check if cracklib was installed correctly
if cracklib == str("n"):
print("Ok, fix it and come back!")
quit()
elif cracklib == str("y"):
print("Proceeding...")
# Tool Removal
print("Going through services and hacking tools commonly used (refer to the readme to see what you do and dont need)... ")
time.sleep(2)
# Apache2 removal
os.system("clear")
apache2 = input("Is apache2 a critical service or do you need it? y/n ")
if apache2 == str("y"):
print("Leaving apache2 installed!")
elif apache2 == str("n"):
print("Removing apache2 and all dependencies completley... ")
os.system("sudo apt autoremove apache2 -y")
# Wireshark removal
os.system("clear")
wireshark = input("Is wireshark a critical service or do you need it? y/n ")
if wireshark == str("y"):
print("Leaving wireshark installed!")
elif wireshark == str("n"):
print("Removing wireshark and all dependencies completley... ")
os.system("sudo apt autoremove wireshark -y")
# Ophcrack removal
os.system("clear")
ophcrack = input("Is ophcrack a critical service or do you need it? y/n ")
if ophcrack == str("y"):
print("Leaving ophcrack installed!")
elif ophcrack == str("n"):
print("Removing ophcrack and all dependencies completley... ")
os.system("sudo apt autoremove ophcrack -y")
# Nmap removal
os.system("clear")
nmap = input("Is nmap a critical service or do you need it? y/n ")
if nmap == str("y"):
print("Leaving nmap installed!")
elif nmap == str("n"):
print("Removing nmap and all dependencies completley... ")
os.system("sudo apt autoremove nmap -y")
# Zenmap removal
os.system("clear")
zenmap = input("Is zenmap a critical service or do you need it? y/n ")
if zenmap == str("y"):
print("Leaving zenmap installed!")
elif zenmap == str("n"):
print("Removing zenmap and all dependencies completley... ")
os.system("sudo apt autoremove zenmap -y")
# Kismet removal
os.system("clear")
kismet = input("Is kismet a critical service or do you need it? y/n ")
if kismet == str("y"):
print("Leaving kismet installed!")
elif kismet == str("n"):
print("Removing kismet and all dependencies completley... ")
os.system("sudo apt autoremove kismet -y")
# Hashcat removal
os.system("clear")
hashcat = input("Is hashcat a critical service or do you need it? y/n ")
if hashcat == str("y"):
print("Leaving hashcat installed!")
elif hashcat == str("n"):
print("Removing hashcat and all dependencies completley... ")
os.system("sudo apt autoremove hashcat -y")
# Nikto removal
os.system("clear")
nikto = input("Is nikto a critical service or do you need it? y/n ")
if nikto == str("y"):
print("Leaving nikto installed!")
elif nikto == str("n"):
print("Removing nikto and all dependencies completley... ")
os.system("sudo apt autoremove nikto -y")
# Hydra removal
os.system("clear")
hydra = input("Is hydra a critical service or do you need it? y/n ")
if hydra == str("y"):
print("Leaving hydra installed!")
elif hydra == str("n"):
print("Removing hydra and all dependencies completley... ")
os.system("sudo apt autoremove hydra -y")
# Netcat removal
os.system("clear")
netcat = input("Is netcat a critical service or do you need it? y/n ")
if netcat == str("y"):
print("Leaving netcat installed!")
elif netcat == str("n"):
print("Removing netcat and all dependencies completley... ")
os.system("sudo apt autoremove netcat -y")
# Going through packages manually
print(bcolors.GREEN + """Synaptic is going to open, click the button that says 'status' then click installed and check all those packages
You're looking for hacking tools and other random stuff, when you find stuff mark it and when done apply it
When you're done, make sure to exit synaptic for the script to continue. DOCUMENT WHAT YOU DELETE SOMEWHERE""" + bcolors.NORMAL)
time.sleep(7)
os.system("sudo synaptic")
time.sleep(3)
os.system("clear")
# Firewall settings
os.system("sudo ufw enable")
print(bcolors.YELLOW + "Set it to office environment and make sure to set incoming to reject it not deny" + bcolors.NORMAL)
time.sleep(3)
os.system("sudo gufw")
os.system("clear")
# Runs a clamscan
os.system("clear")
print("Running AV scan")
time.sleep(1)
os.system("sudo clamscan")
time.sleep(4)
os.system("clear")
# rkhunter thing
print("This is going to go through and check for rootkits/bootkits")
time.sleep(2)
rkhunter = input("Rkhunter scan if this is the first time you run this script type y if it isnt type n")
if rkhunter == str("y"):
os.system("sudo rkhunter --check")
elif rkhunter == str("n"):
print("Skipping rkhunter scan")
time.sleep(1)
os.system("clear")
# PAM Auth Setup
# Doing PAM Password settings
pamq = input(bcolors.GREEN + "Do you need to do pam? y/n" + bcolors.NORMAL)
if pamq == str("y"):
print(bcolors.RED + "This is going to do PAM Password Complexity" + bcolors.NORMAL)
os.system("sudo cat common-password > /etc/pam.d/common-password")
print("Finished doing PAM Auth... doing pam password complexity...")
os.system("clear")
time.sleep(3)
print(bcolors.YELLOW + "Doing PAM remember AUTH PASSWORD settings" + bcolors.NORMAL)
os.system("sudo echo 'auth required pam_tally2.so deny=5 onerr=fail unlock_time=1800' >> /etc/pam.d/common-auth")
os.system("clear")
print(bcolors.YELLOW + "Done with PAM" + bcolors.NORMAL)
elif pamq == str("n"):
print("Skipping PAM settings...")
# Doing file permissions on files
print(bcolors.GREEN + "Doing file permissions for: /etc/passwd /etc/shadow" + bcolors.NORMAL)
os.system("sudo chmod 600 /etc/shadow")
os.system("sudo chmod 644 /etc/passwd")
os.system("clear")
# Check for crontabs
print(bcolors.GREEN + "Go check all these located in /var/spool, comment or delete anything that isnt supposed to be startup!" + bcolors.NORMAL)
os.system("sudo ls /var/spool")
time.sleep(7)
os.system("clear")
# Sodoers check
print(bcolors.GREEN + "Check these files in the following directory using another terminal: /etc/sudoers.d" + bcolors.NORMAL)
time.sleep(4)
os.system("sudo ls /etc/sudoers.d")
os.system("clear")
# Visudo
print(bcolors.YELLOW + "Make sure only root is the only one" + bcolors.NORMAL)
time.sleep(4)
os.system("sudo visudo")
os.system("clear")
# END SEQUENCE print("Finished! A list of what this tool has completed will be sent to a folder")
| [
"noreply@github.com"
] | noreply@github.com |
73ae1bcec05d76e0187dedc15f40138e3430f48b | d95026fe7dbcf749b3cceb3cfd35ffed52d8bc88 | /manytomanybug/urls.py | 634882c5c914a063fdda9707b828634be91ead9c | [] | no_license | arthurio/manytomanybug | c9d3b028862dcef1cbe56d1b7dfb4c437169bdea | c90bc0451e8ed576a0a52933f3436a7a71c34426 | refs/heads/master | 2020-12-24T14:18:43.901741 | 2015-06-18T09:24:02 | 2015-06-18T09:24:02 | 37,646,303 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 763 | py | """manytomanybug URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.8/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Add an import: from blog import urls as blog_urls
2. Add a URL to urlpatterns: url(r'^blog/', include(blog_urls))
"""
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
]
| [
"arthur@punchtab.com"
] | arthur@punchtab.com |
2cae6553ffc000be3d826179dda5c3915a8113fa | 36c0890762bde19b7ebb6330da79da46b3c2332e | /course/9/9-3.py | 59da17882606ae2032b5175ee8c9a4918636fc4f | [] | no_license | chuanyedadiao/Python-Practice | 143857630d8f3129ff62d1063c2b84334dbcd5e7 | a5322241b38f2d1304c2a56ee28f4a2ae12490e2 | refs/heads/master | 2021-01-06T14:45:52.935151 | 2020-03-03T06:47:20 | 2020-03-03T06:47:20 | 241,365,934 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 642 | py | class User():
def __init__(self,first_name,last_name,age):
self.first_name=first_name
self.last_name = last_name
self.age=age
def describe_user(self):
print("FirstName:\t"+self.first_name)
print("LastName:\t"+self.last_name)
print("Age:\t\t"+str(self.age))
def greet_user(self):
print("Ohiyo Hi "+self.first_name)
person1 = User("Yu Shun","Xu",21)
person1.describe_user()
person1.greet_user()
print()
person1 = User("Xiang","Zhou",20)
person1.describe_user()
person1.greet_user()
print()
person1 = User("Yi Yue","Zhong",21)
person1.describe_user()
person1.greet_user()
| [
"chuanyedadiao@gmail.com"
] | chuanyedadiao@gmail.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.