blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 2 616 | content_id stringlengths 40 40 | detected_licenses listlengths 0 69 | license_type stringclasses 2 values | repo_name stringlengths 5 118 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringlengths 4 63 | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 2.91k 686M ⌀ | star_events_count int64 0 209k | fork_events_count int64 0 110k | gha_license_id stringclasses 23 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 220 values | src_encoding stringclasses 30 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 2 10.3M | extension stringclasses 257 values | content stringlengths 2 10.3M | authors listlengths 1 1 | author_id stringlengths 0 212 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
cac26fe53126cbc8d9f1890a5eb47aa89c59b35b | d998988cb9f6f73be4f5b3cc302424119f66c856 | /MGITGatePass/basic_app/migrations/0005_auto_20191217_2052.py | 78882130aea3599e56d61a033f94c7640406fed3 | [] | no_license | sai-sambhu/MGITGatePass | cab9ea545f8b8c5fd959613b364dd161c416fab5 | a14ad7fb79163c1c53d81a5732d2ac0e1ec5544e | refs/heads/master | 2020-11-26T00:51:49.287696 | 2019-12-18T20:02:49 | 2019-12-18T20:02:49 | 228,912,577 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 645 | py | # Generated by Django 2.2.5 on 2019-12-17 15:22
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('basic_app', '0004_auto_20191217_2031'),
]
operations = [
migrations.AddField(
model_name='studentprofileinfo',
name='roll',
field=models.CharField(default='17261A0551', max_length=10),
),
migrations.AlterField(
model_name='studentprofileinfo',
name='profile_pic',
field=models.ImageField(blank=True, default='default.jpeg', upload_to='profile_pics_students'),
),
]
| [
"saisambhuprasad@gmail.com"
] | saisambhuprasad@gmail.com |
ee1a31f88eeb3c7e9f45e9d6e74e4f4ac8581dbf | 53fab060fa262e5d5026e0807d93c75fb81e67b9 | /backup/user_381/ch15_2020_09_14_14_10_44_836878.py | 03149e67069cc3803fab0866de1f386bfbe66feb | [] | no_license | gabriellaec/desoft-analise-exercicios | b77c6999424c5ce7e44086a12589a0ad43d6adca | 01940ab0897aa6005764fc220b900e4d6161d36b | refs/heads/main | 2023-01-31T17:19:42.050628 | 2020-12-16T05:21:31 | 2020-12-16T05:21:31 | 306,735,108 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 177 | py | def chris(nome):
if chris == nome:
return 'Todo mundo odeia o Chris'
else:
return 'Olá, {0}'.format(nome)
nome = input('Qual seu nome?')
| [
"you@example.com"
] | you@example.com |
2e89c56bda2683eb9350f7721d4902cb212adae9 | 0bd2de354c696939fedcecb72c924eaa95feb5b2 | /job/views.py | 41186e674e7661a4fa1918f1f8466c05446a4a6b | [] | no_license | RishabhVerma098/personel-portfolio | 94a0643e522d8c43082712b8c0ec5f9094abebc8 | c11aca5b878a5d7490c7159f3b447a45013a5d04 | refs/heads/master | 2021-07-17T09:45:38.546492 | 2019-02-09T14:28:20 | 2019-02-09T14:28:20 | 169,574,563 | 0 | 0 | null | 2020-06-07T14:32:31 | 2019-02-07T13:18:29 | HTML | UTF-8 | Python | false | false | 190 | py | from django.shortcuts import render
# Create your views here.
from .models import jobs
def home(request):
job = jobs.objects
return render(request, 'job/home.html', {'jobs': job})
| [
"vermarishabh0987@gmail.com"
] | vermarishabh0987@gmail.com |
8bdf10e79cd099752ace442f1093c5d4cfb72eca | a985414206b78dd6e15b485f1881287da6f67895 | /app/__init__.py | 4ad6d80e134bc4d66d4ae9840d13c461da368b7e | [] | no_license | arangurenalonso/hackathon_sem11 | 927873c0010e02f5247b47eae2cbc882b3f03042 | 1d2810871f491e96cd414390234af9f6f560421f | refs/heads/main | 2023-03-22T00:47:33.853778 | 2021-03-12T21:10:20 | 2021-03-12T21:10:20 | 347,198,087 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,165 | py | from flask import Flask
from pathlib import Path
from config import Config
from flask_restx import Api
from app.category.categoryResource import CategoryResource, CategoriesResource, category_ns, categories_ns
from app.producto.productoResource import ProductResource, producto_ns
app = Flask(__name__)
app.config.from_object(Config)
authorizations = {
'Bearer Auth': {
'type': 'apiKey',
'in': 'header',
'name': 'Authorization'
}
}
api = Api(app,
title='Pachaqtec Blog',
version='v1',
description='RESTApi Blog',
prefix='/api/', doc='/swagger/',
contact='Jeancarlos De la cruz',
security='Bearer Auth',
authorizations=authorizations,
contact_url='https://www.linkedin.com/in/jeancarlosdelacruz/')
api.add_namespace(category_ns)
category_ns.add_resource(CategoryResource, '/<int:id>')
api.add_namespace(categories_ns)
categories_ns.add_resource(CategoriesResource, '')
api.add_namespace(producto_ns)
producto_ns.add_resource(ProductResource,'')
from app.category import categoryModel
from app.producto import productoModel
| [
"aranguren.alonso@gmail.com"
] | aranguren.alonso@gmail.com |
191a36449075d84cb10df8758f93f23b02e0ef64 | cfb33b4a1216387c8417af9d04058b87e5d557a9 | /Neutrons/Fast/1SourceMLEM.py | 0d28aff7eb59e63f5c07e39fb4835a1fb576910b | [] | no_license | loomisdevon/DRRSMask | 8c3428acbab7095e8df29008d73d5ab4e7d17b0c | 181955f68b0043a02b38e14b29654779a220b40a | refs/heads/master | 2023-01-15T03:46:43.478169 | 2020-11-26T06:37:36 | 2020-11-26T06:37:36 | 277,006,493 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,054 | py | # Dual Rotating Radiation Scattering Mask MCNP Code
# Rotate neutron and gamma flux from 0 to 2PI around mask and return signal received from detector
# Authors: Ivan Novikov and Devon Loomis
import os
import os.path
import math
import fileinput
import shutil
import subprocess
import io
import string
import numpy as np
import matplotlib.pyplot as plt
import time
import math
from decimal import Decimal
import csv
from tqdm import *
########### GLOBALS #############
#Configuration Name (this is name of input file without .txt)
CONFIG_NAME = 'DDRS3_rand2_absorber1Source'
SOURCE_NAME = 'source3'
tMATRIXCONFIG_NAME = 'DDRS3_rand2_absorber'
tMatrixFilename = tMATRIXCONFIG_NAME + "tMatrix.csv"
#####################
# Source Info
R = 100
Phi = 45
Theta = 140
##############
#################################
###################################### creating all MCNP input files for simulation of the increasing relative distance between source and detector #######################
''' Parameters:
file_name: MCNP input template
y-pos: y position of center of circle of source path
r: radius of circle of source path
init: inital angle away from xy-plane
final angle away from xy-plane
angle step size
limit: longest distance between the two
'''
def smoothing(fluxArray, smoothingParameter):
smoothingArray = []
for i in range(len(fluxArray)):
totSum = 0
numPnts = 0
if (i - smoothingParameter < 0):
for j in range(i,0,-1):
totSum += fluxArray[j]
numPnts += 1
for k in range(i,i+smoothingParameter,1):
totSum += fluxArray[k]
numPnts += 1
elif (i + smoothingParameter > len(fluxArray)):
for j in range(i,len(fluxArray),1):
totSum += fluxArray[j]
numPnts += 1
for k in range(i,i-smoothingParameter,-1):
totSum += fluxArray[k]
numPnts += 1
else:
for j in range(i,i+smoothingParameter,1):
totSum += fluxArray[j]
numPnts += 1
for k in range(i,i-smoothingParameter,-1):
totSum += fluxArray[k]
numPnts += 1
average = totSum/numPnts
smoothingArray.append(average)
return smoothingArray
def createFiles(file_name, phi, rad, init, final, step_size):
fileList = []
marker=0
rad_phi = math.radians(phi)
for new_theta in range(init, final, step_size):
text_search = None
f =open(file_name)
for line in f:
words = line
sdef = words[0:4]
if (sdef == "SDEF"):
text_search = words
break
f.close()
rad_theta = math.radians(new_theta)
x_pos = round(rad * np.cos(rad_theta)*np.sin(rad_phi),3)
y_pos = round(rad * np.sin(rad_theta)*np.sin(rad_phi),3)
z_pos = round(rad * np.cos(rad_phi),3)
r_mag = np.sqrt(x_pos**2+y_pos**2+z_pos**2)
vecx_pos = round(-x_pos/r_mag,3)
vecy_pos = round(-y_pos/r_mag,3)
vecz_pos = round(-z_pos/r_mag,3)
#theta_rad = np.arctan(z_pos/r)
#vecz_pos = round(-1 * (theta_rad/(np.pi/2)),5)
#replacement_text = sdef + " ERG = 1.42 POS " + str(x_pos) + " " + str(y_pos) + " " + str(z_pos) + " VEC= " + str(vecx_pos) + " " + str(vecy_pos) + " " + str(vecz_pos) + " DIR=d1 par=n" + "\n"
replacement_text = sdef + " ERG = 2.0 POS " + str(x_pos) + " " + str(y_pos) + " " + str(z_pos) + " VEC= " + str(vecx_pos) + " " + str(vecy_pos) + " " + str(vecz_pos) + " DIR=d1 WGT 20 par=n" + "\n"
#replacement_text = sdef + " ERG = 1.42 POS " + str(x_pos) + " " + str(y_pos) + " " + str(z_pos) + " par=n" + "\n"
read_name = file_name
write_name = CONFIG_NAME + SOURCE_NAME + "_" + str(new_theta) + ".txt"
f1 = open(read_name, 'r')
f2 = open(write_name, 'w')
for lines in f1:
f2.write(lines.replace(text_search, replacement_text))
f1.close()
f2.close()
fileList.append(write_name)
return (fileList)
################################# delete runtpe files after every set of commands and delete all output files and input files after program run #######################
''' Parameters
directory: directory containing all files
file: KSEF_2 #####################
remove_all: test to determine whether to delete all files or only runtpe files
'''
def removeFiles(directory, file1, file2, file3, outfile, initfile, t_file, save_one, remove_all):
dir_name = directory
for fname in os.listdir(dir_name):
if (fname != initfile and fname != t_file):
if fname.startswith("binRun"):
os.remove(os.path.join(dir_name, fname))
if (fname.startswith(file1[:-4]) or fname.startswith(outfile[:-4])) and remove_all:
if (fname != file1):
os.remove(os.path.join(dir_name, fname))
if (fname.startswith(file1[:-4]) or fname.startswith(outfile[:-4])) and save_one:
if (fname != file1 and fname != file2 and fname != file3):
os.remove(os.path.join(dir_name, fname))
####################### read MCNP output file, find and return flux value #########################
#######################_file_: MCNP output file name ##################################
'''
def readFlux(_file_):
flux_ = 0
error_ = 0
with open(_file_, 'r') as outfile:
for line in outfile:
if ('+ *Gamma flux in detector*' in line):
lines = [outfile.readline() for i in range(9)] #this reads 9 lines after the fc4 comment
spectrum = [outfile.readline() for i in range(13)] #this reads 13 lines which contain spectrum
#each line has an index [0]-[12]
#print(type(spectrum[1]))
#print(spectrum[1])
#print(float(spectrum[1].split()[1])) #this splits spectrum[i] using spaces
#each spectrum[i].split() has three new indeces [0]-[2]
#float converts each string to float
#Neutron energy is in [0]
#Neutron counts are in [1]
#Error is in [2]
#tmp = 0.0
#print (spectrum)
for j in range(13):
flux_ += float(spectrum[j].split()[1])
error_ += float(spectrum[j].split()[2])
#Fluxin3[i] = tmp
return flux_, error_
'''
def readFlux(_file_,energyBin, binWrite):
flux_Arr = []
error_Arr = []
flux_ = 0
error_ = 0
with open(_file_, 'r') as outfile:
for line in outfile:
if ('+ *Neutron Flux In Detector*' in line):
lines = [outfile.readline() for i in range(9)] #this reads 9 lines after the fc4 comment
spectrum = [outfile.readline() for i in range(energyBin+1)] #this reads 13 lines which contain spectrum
#each line has an index [0]-[12]
#print(type(spectrum[1]))
#print(spectrum[1])
#print(float(spectrum[1].split()[1])) #this splits spectrum[i] using spaces
#each spectrum[i].split() has three new indeces [0]-[2]
#float converts each string to float
#Neutron energy is in [0]
#Neutron counts are in [1]
#Error is in [2]
#tmp = 0.0
for j in range(energyBin+1):
if (binWrite == j and binWrite != 0):
flux_ = float(spectrum[j].split()[1])
error_ = float(spectrum[j].split()[1])
#print (float(spectrum[j].split()[1]))
#flux_Arr.append(float(spectrum[j].split()[1]))
#error_Arr.append(float(spectrum[j].split()[2]))
#flux_ += float(spectrum[j].split()[1])
#error_ += float(spectrum[j].split()[2])
if (binWrite == 0):
flux_ = float(spectrum[energyBin].split()[1])
error_ = float(spectrum[energyBin].split()[2])
#Fluxin3[i] = tmp
return flux_, error_
def initialize(_file_):
global intensity, activity, nps, t
global radius, init_theta, final_theta, step_theta
global init_phi, final_phi, step_phi
global packet
with open(_file_,"r", newline='') as file:
file.readline()
intensity = float(file.readline()[12:])
activity = float(file.readline()[11:])
nps = float(file.readline()[6:])
t = float(file.readline()[4:])
file.readline()
radius = float(file.readline()[9:])
init_theta = int(file.readline()[13:])
final_theta = int(file.readline()[14:])
step_theta = int(file.readline()[13:])
init_phi = float(file.readline()[11:])
final_phi = float(file.readline()[12:])
step_phi = float(file.readline()[11:])
file.readline()
file.readline()
packet = int(file.readline()[9:])
#**********************MAIN**************************
#dir_ = 'C:\\Users\\devon\\Documents\\DRRSMask\\Working_Version\\MLEM\\'
dir_ = os.path.dirname(os.path.abspath(CONFIG_NAME+SOURCE_NAME)) + "\\"
file_ = CONFIG_NAME + SOURCE_NAME + '.txt'
outFile_ = CONFIG_NAME + SOURCE_NAME + '_out.txt'
file_name_ = dir_ + file_
outFile_name_ = dir_ + outFile_
keepInFile = CONFIG_NAME + SOURCE_NAME + '_0.txt'
keepOutFile = CONFIG_NAME + SOURCE_NAME + '_out0.txt'
init_file= dir_ + 'init.txt'
t_file = CONFIG_NAME + "tMatrix.csv"
intensity, activity, nps, t = 0,0,0,0
radius,init_theta,final_theta,step_theta = 0,0,0,0
init_phi,final_phi,step_phi = 0,0,0
packet = 0
initialize(init_file)
originalThetaCountsArray = []
originalThetaErrorArray = []
init_theta = Theta
final_theta = init_theta+360
transmissionMatrix = []
start = time.time()
removeFiles(dir_, file_, keepInFile, keepOutFile, outFile_, init_file, t_file, False, True) # purge directory of any existing MCNP files from previous run
#files = createFiles(file_name_, z, radius, init_ang, final_ang, step)
files = createFiles(file_name_, Phi, R, init_theta, final_theta, step_theta) # create all MCNP input files
commands = []
outFileList = []
j = init_theta
#create set of commands for subprocess of all input files
for i in range(int((final_theta - init_theta) / step_theta)):
binFile = "binRun" + str(j) + ".r"
outFile = (CONFIG_NAME + SOURCE_NAME + "_out" + str(j) + ".txt")
commands.append("mcnp6 i=" + files[i] + " o=" + outFile + " runtpe=" + binFile)
outFileList.append(outFile)
j += step_theta
print("Simulating...")
# give subprocess pak amount of parallel programs to execute until all commands are executed
for x in tqdm(range(0,int((final_theta - init_theta) / step_theta),(packet))):
if (x < (len(commands) - packet)):
commandsub = commands[x:(x+packet)]
else:
commandsub = commands[x:]
processes = [subprocess.Popen(cmd, shell=True, stdout=subprocess.PIPE, cwd=dir_) for cmd in commandsub]
removeFiles(dir_, file_, keepInFile, keepOutFile, outFile_, init_file, t_file, False, False) # remove runtpe files
for p in processes:
p.wait()
print ("Checkpoint")
theta = init_theta
fluxList = []
errorList = []
sourceThetaList = []
#use for neutrons
#energyBinOfInterest = 13
#use for gammas
energyBinOfInterest = 100
############################read and gather flux values and source distances for each output file and add them to lists###################################
for f in outFileList:
flux, error = readFlux(f, energyBinOfInterest,40)
fluxList.append(flux)
errorList.append(error)
rad_theta = math.radians(theta)
sourceThetaList.append(rad_theta)
theta += step_theta
removeFiles(dir_, file_, keepInFile, keepOutFile, outFile_, init_file, t_file, True, False)
end = time.time()
print("Runtime: ", round((end - start)/60, 2), " mins")
rawFluxArray = np.array(fluxList)
#print (rawFluxArray)
fluxArray = np.array(smoothing(fluxList, 8))
#fluxArray = np.array(fluxList)
errorArray = np.array(errorList)
#print (errorArray)
thetaArray = np.array(sourceThetaList)
countsArray = fluxArray * intensity * t
countsSum = np.sum(countsArray)
#normalizedCountsArray = countsArray / countsSum
normalizedCountsArray = np.copy(countsArray)
normalizedCountsErr = np.sqrt((1/countsArray) + (1/countsSum))
normalizedCountsErrorArray = np.multiply(normalizedCountsArray, normalizedCountsErr)
with open(CONFIG_NAME + SOURCE_NAME + "data.csv","w+", newline='') as file:
writer=csv.writer(file,delimiter=',')
for a in normalizedCountsArray:
writer.writerow([a])
with open(CONFIG_NAME + SOURCE_NAME + "background.csv","w+", newline='') as file:
writer=csv.writer(file,delimiter=',')
for b in normalizedCountsErrorArray:
writer.writerow([b])
###########################END MAIN############################### | [
"devon.loomis@scientic.com"
] | devon.loomis@scientic.com |
5ab61d683509ed50cb0997df3f6350787fa30c58 | e30afd7f5166afed9ebfd740765d74a152175f54 | /algorithm-40case/leetcode_exercise/add_two_numbers.py | 9851649b651ca875717c0399a4c9ef6a092d3e53 | [] | no_license | wqdchn/geektime | 214c4ee627023723585ddbc66f3066f816a4fc76 | d953f631d93a4493aed26392954ca1f18a440938 | refs/heads/master | 2021-12-27T16:19:03.680654 | 2021-12-26T11:21:35 | 2021-12-26T11:21:35 | 174,132,464 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 977 | py | # @program: PyDemo
# @description: https://leetcode.com/problems/add-two-numbers/
# @author: wqdong
# @create: 2019-10-01 14:31
class ListNode:
def __init__(self, x):
self.val = x
self.next = None
class Solution:
def addTwoNumbers(self, l1: ListNode, l2: ListNode):
divmod_carry = 0
res = curr = ListNode(0)
while l1 or l2 or divmod_carry:
if l1:
divmod_carry += l1.val
l1 = l1.next
if l2:
divmod_carry += l2.val
l2 = l2.next
divmod_carry, divmod_val = divmod(divmod_carry, 10)
curr.next = curr = ListNode(divmod_val)
return res.next
s = Solution()
l1 = ListNode(2)
l1.next = ListNode(4)
l1.next.next = ListNode(3)
l2 = ListNode(5)
l2.next = ListNode(6)
l2.next.next = ListNode(4)
res = s.addTwoNumbers(l1, l2)
while res:
print(res.val)
res = res.next
| [
"wqdong.chn@gmail.com"
] | wqdong.chn@gmail.com |
af6c43ee70e8b9d1ae987c97a80ae8707f4b001e | 59dbbdf5d29d2490ec8a697dc137aa7456479e89 | /usage/meta.py | 492a6b0640a3b1458ec28a2c5f9d8bdf040928ea | [
"Apache-2.0"
] | permissive | absalon-james/usage | 15d424599528bec7d3184a72b5e9754c325e46ed | a67ceddda8a14244526b3b3a40c0c3feec7035d2 | refs/heads/master | 2021-01-21T14:58:06.023114 | 2016-10-03T20:56:16 | 2016-10-03T20:57:46 | 57,158,746 | 1 | 1 | null | null | null | null | UTF-8 | Python | false | false | 96 | py | version = '0.1.2'
description = "Python tool for collecting usage information from ceilometer."
| [
"james.absalon@rackspace.com"
] | james.absalon@rackspace.com |
975d016f867e515a5d0d42fecfb3e22ccc3a61ff | 55ddcae82338890a7101b2ff6db0856463702314 | /perfectcushion/shop/admin.py | 1ac35dfa7e64e770e7c84c4021ff8ced72dc2336 | [] | no_license | rixinhaha/DjangoEcommerce | d31d6e8c7a4a40ba3f32d0e27ef203c59475c1dc | 0e3a188e8276bbfb63901747f553dd2ab483c284 | refs/heads/master | 2020-08-03T21:18:14.750498 | 2019-09-30T15:30:09 | 2019-09-30T15:30:09 | 211,887,708 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 546 | py | from django.contrib import admin
from .models import Category,Product
# Register your models here.
class CategoryAdmin(admin.ModelAdmin):
list_display = ['name', 'slug']
prepopulated_fields = {'slug':['name',]}
admin.site.register(Category, CategoryAdmin)
class ProductAdmin(admin.ModelAdmin):
list_display= ['name', 'price', 'stock', 'available', 'created', 'updated']
list_editable= ['price', 'stock', 'available']
prepopulated_fields = {'slug':['name',]}
list_per_page = 20
admin.site.register(Product,ProductAdmin)
| [
"rixinhaha@gmail.com"
] | rixinhaha@gmail.com |
41e2a093e82ce2eb956957ab8d320d97248524fb | 1a1a10576c4fabe2879feb00393fe8d3f5211d9b | /todo/settings.py | dfb08b572fb8286e00d5acfcf40919ff7f4a9f4c | [] | no_license | Machele-codez/todo-by-machele | 8a50e7e6f9e560ba40dd7816f1c34324cf4c4f8d | f96903f24a5fdfb587cbf04977b0c2aaf5a92879 | refs/heads/master | 2022-09-13T02:42:36.410948 | 2020-06-03T19:59:33 | 2020-06-03T19:59:33 | 269,172,154 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,336 | py | """
Django settings for todo project.
Generated by 'django-admin startproject' using Django 2.2.7.
For more information on this file, see
https://docs.djangoproject.com/en/2.2/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/2.2/ref/settings/
"""
import os
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/2.2/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'gk(c+=qm^h6-b-)g%=ej0%kzmgnbwl=k^35uupjpspq)ql^uw2'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'django.contrib.humanize',
'apps.accounts',
'apps.tasks',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'todo.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [os.path.join(BASE_DIR, 'templates')],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'todo.wsgi.application'
# Database
# https://docs.djangoproject.com/en/2.2/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),
}
}
# Password validation
# https://docs.djangoproject.com/en/2.2/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/2.2/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/2.2/howto/static-files/
STATIC_URL = '/static/'
STATICFILES_DIRS = [os.path.join(BASE_DIR, 'static')]
STATIC_ROOT = os.path.join(BASE_DIR, 'staticfiles')
LOGIN_REDIRECT_URL = 'tasks:all_tasks ' | [
"smithbeblack@gmail.com"
] | smithbeblack@gmail.com |
3cd6e2099d0754d1a415a93ab25595f0ada97a68 | 77909a8a93f60759e0fd32fb632d937c7c8d4d68 | /curso_em_video_exercises/desafio10.py | c1c5bc683719ab4335f0251c26bf81cd73603c19 | [] | no_license | euricoteles/python | 13e39bf0b5916b69794dac39dc55a213b5443718 | dae10d87a9923646dd8257a2ce3da91dc355b603 | refs/heads/master | 2021-09-04T14:51:47.147309 | 2018-01-19T16:57:27 | 2018-01-19T16:57:27 | 116,609,339 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 286 | py | # Crie um script que leia quanto dinheiro uma pessoa tem na carteira e converter em dolares.
# variables received by input
valor = int(input('Qual o valor para converter em dolares:'))
dolares = (1.18*valor)/1
# print information
print('O valor em dolares fica : {}'.format(dolares)) | [
"euriconaz@hotmail.com"
] | euriconaz@hotmail.com |
dfd3789007df10b47fed17eb4ecef1dbbe054537 | 2edbdd6763f86aca4f6ee67ced390fb477ed0e44 | /udf/extract_type_modify_law.py | b2453cfaf4db58be6073271509e34dc2f95f64b6 | [] | no_license | nhatan172/deepdive | f498901c3faa474d3ab5166ef7a694e1626e2f01 | 492afee641436e4d5a068a7cec1ff3969d964518 | refs/heads/master | 2021-04-06T02:05:17.095127 | 2018-04-12T04:07:45 | 2018-04-12T04:07:45 | 124,988,988 | 0 | 0 | null | 2018-03-21T23:19:06 | 2018-03-13T03:49:20 | Python | UTF-8 | Python | false | false | 12,317 | py | #!/usr/bin/env python
# -*- coding:utf8 -*-
from deepdive import *
import re
import handle_string
import divlaw
def lenIterator(list):
sum = 0
for i in list :
sum += 1
return sum
def getTitle(string):
temp = re.finditer(r"\:(\s|\n|\*|\_|\#)*(\“|\")",string,re.DOTALL)
end_title = len(string)
if lenIterator(temp) > 0 :
temp = re.finditer(r"\:(\s|\n|\*|\_|\#)*(\“|\")",string,re.DOTALL)
for i in temp:
end_title = i.start()
break
return string[:end_title]
def get_numerical_symbol(title):
title = re.sub(r'(\“(.(?!\“|\”))+.{2})|(\"(.(?!\"))+.{2})',"",title,re.M|re.DOTALL)
get_title1 = re.search(r'(của\s.*)\s(đã được|được)',title)
get_title = re.search(r'[0-9]+(/[0-9]+)*((/|-)[A-ZĐƯ]+[0-9]*)+(\s|\_|\#|\*|\.|\\)',title,re.M|re.I)
# get_id = re.search(r'[0-9]+(/[0-9]+)*((/|-)[A-ZĐ]+[0-9]*)+',get_content.group())
# get_title1 = re.search(r'([0-9]+(/[0-9]+)*((/|-)[A-ZĐ]+[0-9]*)\s(đã được))|([0-9]+(/[0-9]+)*((/|-)[A-ZĐ]+[0-9]*)\s(được))',title)
if(get_title1 is not None):
number = re.search(r'[0-9]+(/[0-9]+)*((/|-)[A-ZĐƯ]+[0-9]*)+(\s|\_|\#|\*|\.|\\)',get_title1.group())
if(number is not None):
return (re.search(r'[0-9]+(/[0-9]+)*((/|-)[A-ZĐƯ]+[0-9]*)+',number.group(),re.U|re.I)).group()
elif ((get_title is not None) and (get_title1 is None)):
return (re.search(r'[0-9]+(/[0-9]+)*((/|-)[A-ZĐƯ]+[0-9]*)+',get_title.group(),re.U|re.I)).group()
else :
return None
@tsv_extractor
@returns(lambda
law_id ="text",
type = "int",
doc_content_update = "text",
symbol = "text",
position = "text",
modified_law_date_release = "text"
:[])
def extract(
law_id = "text",
totalLaw = "int",
law_content = "text",
law_len = "int",
totalItem = "int",
item_content = "text",
item_len = "int",
totalpoint = "int",
point_content = "text",
part_index ="int",
chap_index ="int",
sec_index ="int",
law_index ="int",
item_index ="int",
point_index ="int",
numerical_symbol = "text",
date_released ="text"
):
doc_content_update = None
if law_content is not None:
# law_content = handle_string.to_unicode(law_content)
law_content = law_content[:law_len]
# pass
# law_content = law_content.encode('utf-8')
if (item_content is not None) :
# # item_content = handle_string.to_unicode(item_content)
# # if item_len != len(item_content):
item_content = item_content[:item_len]
# pass
# item_content = item_content.encode('utf-8')
number = None
type = 0
point = 0
p = re.compile(r'((((S|s)ửa đổi)(\s|\,)*((b|B)ổ sung)*)|((b|B)ổ sung))')
p1= re.compile(r'(đã\s|đã được\s)((((S|s)ửa đổi)(\s|\,)*((b|B)ổ sung)*)|((b|B)ổ sung))')
position = "0_0_0_0_0_0"
if(totalpoint > 0):
number = get_numerical_symbol(getTitle(point_content))
if(number is not None):
numerical_symbol = number
date_released = None
position = "{}_{}_{}_{}_{}_{}".format(part_index+1,chap_index+1,sec_index+1,law_index+1,item_index+1,point_index+1)
type_modify = re.search(r'(((b|B)ổ sung cụm từ)|((b|B)ổ sung từ))',point_content)
if(type_modify is not None):
type = 3
doc_content_update = point_content
point = 1
else :
type_change_name = re.search(r'(S|s)ửa đổi tên',point_content)
if(type_change_name is not None):
type = 6
doc_content_update = point_content
point = 1
else:
type_delete = re.search(r'(b|B)ãi bỏ',point_content)
inQuote = False
if type_delete is not None :
inQuote = divlaw.itemInQuote(point_content,type_delete.start())
if(type_delete is not None) and not inQuote:
type = 2
doc_content_update = point_content
point = 1
else:
type_delete_text = re.search(r'(((b|B)ỏ cụm từ)|((b|B)ỏ từ))',point_content)
if(type_delete_text is not None):
type = 7
doc_content_update = point_content
point =1
else:
type_add_text = p.finditer(point_content)
type_add_text1 = p1.finditer(point_content)
len1 = lenIterator(type_add_text)
len2 = lenIterator(type_add_text1)
if( (len1 != len2) and (len1 > 0)):
type = 1
doc_content_update = point_content
point = 1
else :
# type_change_text = re.search(r'(t|T)hay\s.*cụm từ',point_content)
type_change_text = re.search(r'((t|T)hay\s)*(cụm\s)*từ\s.*(được\s)*(thay\s)*bằng\s(cụm\s)*từ',point_content)
if(type_change_text is not None):
type = 4
doc_content_update = point_content
point = 1
else :
type_name_to_name = re.search(r'((t|T)ên của\s).+(((S|s)ửa đổi\s)*(\,\s)*((b|B)ổ sung\s)*)(thành)',point_content)
if(type_name_to_name is not None):
type = 5
doc_content_update =point_content
point = 1
else :
point = 0
if(totalItem > 0 and point == 0):
number = get_numerical_symbol(getTitle(item_content))
if(number is not None):
numerical_symbol = number
date_released = None
position = "{}_{}_{}_{}_{}_{}".format(part_index+1,chap_index+1,sec_index+1,law_index+1,item_index+1,0)
type_modify = re.search(r'(b|B)ổ sung cụm từ',item_content)
if(type_modify is not None):
type = 3
doc_content_update = item_content
point = 1
else:
type_change_name = re.search(r'(S|s)ửa đổi tên',item_content)
if(type_change_name is not None):
type = 6
doc_content_update = item_content
point = 1
else:
type_delete = re.search(r'(b|B)ãi bỏ',item_content)
inQuote = False
if type_delete is not None :
inQuote = divlaw.itemInQuote(item_content,type_delete.start())
if(type_delete is not None) and not inQuote:
type = 2
doc_content_update = item_content
point = 1
else:
type_delete_text = re.search(r'(((b|B)ỏ cụm từ)|((b|B)ỏ từ))',item_content)
if(type_delete_text is not None):
type = 7
doc_content_update = item_content
point = 1
else:
# type_add_text = re.search(r'((((S|s)ửa đổi)(\s|\,)*((b|B)ổ sung)*)|((b|B)ổ sung))',item_content)
# if(type_add_text is not None):
type_add_text = p.finditer(item_content)
type_add_text1 = p1.finditer(item_content)
len1 = lenIterator(type_add_text)
len2 = lenIterator(type_add_text1)
if( (len1 != len2) and (len1 > 0)):
type = 1
doc_content_update = item_content
point=1
else:
# type_change_text = re.search(r'(t|T)hay\s.*cụm từ',item_content)
type_change_text = re.search(r'((t|T)hay\s)*(cụm\s)*từ\s.*(được\s)*(thay\s)*bằng\s(cụm\s)*từ',item_content)
if(type_change_text is not None):
type = 4
doc_content_update = item_content
point = 1
else :
type_name_to_name = re.search(r'((t|T)ên của\s).+(((S|s)ửa đổi\s)*(\,\s)*((b|B)ổ sung\s)*)(thành)',item_content)
if(type_name_to_name is not None):
type = 5
doc_content_update = item_content
point = 1
else :
point = 0
# if(totalpoint > 0 and point == 1 ):
# doc_content_update = point_content
if(totalLaw >0 and point == 0 ):
number = get_numerical_symbol(getTitle(law_content))
if(number is not None):
numerical_symbol = number
date_released = None
position = "{}_{}_{}_{}_{}_{}".format(part_index+1,chap_index+1,sec_index+1,law_index+1,0,0)
type_modify = re.search(r'(b|B)ổ sung cụm từ',law_content)
if(type_modify is not None):
type = 3
doc_content_update = law_content
point = 1
else:
type_change_name = re.search(r'(S|s)ửa đổi tên',law_content)
if(type_change_name is not None):
type = 6
doc_content_update = law_content
point = 1
else:
type_delete = re.search(r'(b|B)ãi bỏ',law_content)
inQuote = False
if type_delete is not None :
inQuote = divlaw.itemInQuote(law_content,type_delete.start())
if(type_delete is not None) and not inQuote:
type = 2
doc_content_update = law_content
point = 1
else:
type_delete_text = re.search(r'(((b|B)ỏ cụm từ)|((b|B)ỏ từ))',law_content)
if(type_delete_text is not None):
type = 7
doc_content_update = law_content
point = 1
else:
type_add_text = p.finditer(law_content)
type_add_text1 = p1.finditer(law_content)
len1 = lenIterator(type_add_text)
len2 = lenIterator(type_add_text1)
if( (len1 != len2) and (len1 > 0)):
type = 1
doc_content_update = law_content
point = 1
else:
type_change_text = re.search(r'((t|T)hay\s)*(cụm\s)*từ\s.*(được\s)*(thay\s)*bằng\s(cụm\s)*từ',law_content)
if(type_change_text is not None):
type = 4
doc_content_update = law_content
point = 1
else :
type_name_to_name = re.search(r'((t|T)ên của\s).+(((S|s)ửa đổi\s)*(\,\s)*((b|B)ổ sung\s)*)(thành)',law_content)
if(type_name_to_name is not None):
type = 5
doc_content_update = law_content
point = 1
else :
point = 0
# if(totalItem > 0):
# doc_content_update = item_content
if(point == 1):
yield[
law_id,
type,
doc_content_update,
numerical_symbol,
position,
date_released
]
| [
"nhatan172@gmail.com"
] | nhatan172@gmail.com |
94f4897b31040b8abc8c30479a440a8e0af48906 | e793abb16a44eff7b48df2c774883d7a469f2005 | /local_code/adam/test.py | 1b238455a5659ce30f9888884b4d2cf00e25d092 | [] | no_license | Junyinghuang/DS4S_group2 | a25f2fe6d31f66452fc03ad02594c9c15a4f0017 | 2a3bcd0fad858e7ef00d69342dbdd47853525b83 | refs/heads/master | 2022-06-20T23:43:11.420048 | 2020-05-12T07:07:56 | 2020-05-12T07:07:56 | 257,669,453 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 797 | py | from local_code.adam.new_parameters import get_new_parameters
def test_parameter_creation():
'''
Test to make sure I don't get any wonky results from my parameter adjustor
'''
sigmas = [1,2,3,4]
initial_guesses = [0,0,0,0]
new_guesses = get_new_parameters(sigmas,initial_guesses)
#It would be surprising if any of these were more than, say, 4 sigma away from the initial guess.
is_more_than_four_sigma_away = [int(abs(guess)>(4*sigma)) for guess, sigma in zip(new_guesses,sigmas)]
assert sum(is_more_than_four_sigma_away)==0,"get_new_parameters yields surprising results."
print('Test of parameter creation complete.')
# I've written no other tests, as all the "visualization" code I've written can be tested simply by
# whether or not a plot shows up. | [
"apkunesh@ucdavis.edu"
] | apkunesh@ucdavis.edu |
739c0ed4a80c4bad6b0788a2f025475c8e864f1c | 3ec9ace491cd5d06b5b998e7e309a13bd86c7126 | /tests/system/conftest.py | 535e9d9a46ea535834835fff3376b8c958e75f58 | [
"Apache-2.0"
] | permissive | Jitsusama/lets-do-dns | 64467664f42df053b535156fc773be7e874d0bf5 | faff4bf45e9a4be438e15afbe5caa249fe1e5210 | refs/heads/master | 2021-01-20T02:38:10.953843 | 2017-07-21T03:08:30 | 2017-07-21T03:08:30 | 89,433,363 | 8 | 0 | Apache-2.0 | 2019-10-02T15:07:28 | 2017-04-26T03:24:48 | Python | UTF-8 | Python | false | false | 1,843 | py | try:
import ConfigParser as configparser
except ImportError:
import configparser
import os
import pytest
from requests import post
@pytest.fixture(autouse=True)
def os_environ_reset():
"""Reset os.environ in between test runs."""
original_env = os.environ.copy()
yield
os.environ.clear()
os.environ.update(original_env)
@pytest.fixture(scope='module')
def test_configuration():
"""Read test configuration from :file:`config.ini` file.
The INI file must have a ``[DEFAULT]`` section containing the following
parameters:
* ``do_api_key``
* ``do_domain``
* ``do_hostname``
"""
file_path = os.path.realpath(__file__)
directory_path = os.path.dirname(file_path)
config_file = '%s/config.ini' % directory_path
config = configparser.ConfigParser()
config.read(config_file)
return config
@pytest.fixture
def create_response(
do_base_uri, do_auth_header, do_domain, do_hostname, request):
return post(
'%s/%s/records' % (do_base_uri, do_domain),
headers=do_auth_header,
json={'type': 'TXT',
'name': do_hostname,
'data': request.function.__name__})
@pytest.fixture()
def do_api_key(test_configuration):
return test_configuration.get('DEFAULT', 'do_api_key')
@pytest.fixture
def do_auth_header(do_api_key):
return {'Authorization': 'Bearer %s' % do_api_key}
@pytest.fixture
def do_base_uri():
return 'https://api.digitalocean.com/v2/domains'
@pytest.fixture
def do_domain(test_configuration):
return test_configuration.get('DEFAULT', 'do_domain')
@pytest.fixture
def do_hostname(test_configuration):
return test_configuration.get('DEFAULT', 'do_hostname')
@pytest.fixture
def do_record_id(create_response):
return create_response.json()['domain_record']['id']
| [
"joel@grrbrr.ca"
] | joel@grrbrr.ca |
6ec1b92adbf29c397050d278e8c4ddd379f0e719 | 874c45e64e28ec63829b22738c3e7744dac1aeb7 | /test/rtt/utils.py | 8edbf72b300db2b646aba48fc6af9dc0080c5cc6 | [] | no_license | zenokoller/rtt-timestamp-vs-spinbit | 75372d04a29dc93c6161d61516d0cdb72e684bfc | a2aedc47dd8c48cdf9771e8c747dff422f97fa31 | refs/heads/master | 2020-03-20T18:32:38.876614 | 2018-07-06T15:39:14 | 2018-07-06T15:39:14 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 142 | py | import pandas as pd
def load_dataframe(path: str, key: str) -> pd.DataFrame:
with pd.HDFStore(path) as store:
return store[key]
| [
"zeno.koller@gmail.com"
] | zeno.koller@gmail.com |
0aca43944e8543cb2dde41c31bfa4b1db2c4dc93 | 969be4b7959617a4def52267595ed22a67caeaaa | /wsgi/closetBackend/Invetory/views.py | f0b1837c8705c2e6196fc3f5e170a4867ea44c0a | [] | no_license | pjryan93/closet | 0589d493958e5f9a63c760ebc4c52588622da913 | 04dcf33f0991c547cc27b7214a81ab0d3f149ff3 | refs/heads/master | 2021-01-10T06:15:36.359576 | 2016-03-20T22:02:52 | 2016-03-20T22:02:52 | 54,337,969 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,952 | py | from django.shortcuts import render
from django.http import HttpResponse, HttpResponseRedirect
from django.views.generic import TemplateView
from django.template.context_processors import csrf
from django.shortcuts import render_to_response
from django.template import Context
from django.template.loader import get_template
from django.contrib.auth import authenticate, login
from django.contrib.auth.models import User
from django.contrib.auth import logout
from django.views.decorators.csrf import csrf_protect
from django.contrib.auth.decorators import user_passes_test
from django.conf import settings
from django.utils.decorators import method_decorator
from django.contrib.auth.decorators import login_required
from rest_framework.permissions import IsAuthenticated
from rest_framework.authentication import TokenAuthentication
from rest_framework.views import APIView
from rest_framework.response import Response
from rest_framework.authtoken.models import Token
from models import *
from serializers import *
from rest_framework.renderers import JSONRenderer
import base64, uuid
import cStringIO
import sys
from django.core.files.base import ContentFile
from base64 import b64decode
from django.core.files.images import ImageFile
class HomeView(APIView):
authentication_classes = (TokenAuthentication,)
def post(self,request,format=None):
return Response({'detail': "I suppose you are authenticated"})
def get(self,request,format=None):
print request.user
return Response({'detail': "I suppose you are authenticated"})
class CreateCloset(APIView):
authentication_classes = (TokenAuthentication,)
def get(self,request,format=None):
closets = Closet.objects.filter(owner=request.user)
serializer = ClosetSerializer(closets, many=True)
json = JSONRenderer().render(serializer.data)
defaults = DefaultAmounts.objects.filter(current_closet=closets[0].id)
serializer = ClosetDefaultsSerializer(defaults,many=False)
json_defaults = JSONRenderer().render(serializer.data)
print 'here'
return Response({'closets': json,'defaults':json_defaults})
def post(self,request,format=None):
user = request.user
print request.data
print request.data['name']
print 'done'
if 'name' in request.data and 'gender' in request.data and 'age' in request.data:
closetName = request.data['name']
age = request.data['age']
gender = request.data['gender']
cleaned_gender= 'Male'
if gender == "Male":
cleaned_gender = "M"
elif gender == "Female":
cleaned_gender = "F"
if 'closet_id' in request.data:
current_closet = Closet.objects.get(id = request.data["closet_id"])
current_closet.name = closetName
current_closet.age = age
current_closet.sex = gender
current_closet.save()
return Response({'success': "updated",'name':closetName,'id':current_closet.id})
elif Closet.objects.filter(owner=request.user,name = closetName).count() == 0:
new_closet = Closet(owner = request.user,name = closetName,age = age,sex = cleaned_gender)
new_closet.save()
defaults = DefaultAmounts(current_closet =new_closet )
defaults.save()
sizes = DefaultSizes(current_closet=new_closet)
sizes.setAll(age)
print 'created'
return Response({'success': "created",'name':closetName,'id':new_closet.id})
else:
return Response({'failure':'You have a closet with this name'})
return Response({'failure':'not created'})
class ClosetItem(APIView):
authentication_classes = (TokenAuthentication,)
def get(self,request,format=None):
print request.body
print request.user
closets = Closet.objects.filter(owner=request.user)
if len(closets) == 0:
defaults = DefaultAmounts()
serializer = ClosetDefaultsSerializer(defaults,many=False)
json = JSONRenderer().render(serializer.data)
return Response({'closets': json,'message':'no closets'})
else:
return self.getResponseNoId(request)
if 'id' in request.GET:
closets = Closet.objects.filter(owner=request.user)[0]
serializer1 = ClosetSerializer(closets, many=False)
json_defaults = JSONRenderer().render(serializer1.data)
defaults = DefaultAmounts.objects.filter(current_closet=closets.id)
serializer = ClosetDefaultsSerializer(defaults,many=True)
json = JSONRenderer().render(serializer.data)
clothing_items = ClothingItem.objects.filter(current_closet=closets.id)
item_serializer = ItemSerializer(clothing_items,many=True)
item_json = JSONRenderer().render(item_serializer.data)
size_defaults= DefaultSizes.objects.filter(current_closet=closets.id)
if len(size_defaults) == 0:
default_size = DefaultSizes(current_closet = closets)
default_size .save()
size_defaults= DefaultSizes.objects.filter(current_closet=closets.id)
size_serializer = ItemSizeSerializer(size_defaults,many=True)
size_json = JSONRenderer().render(size_serializer.data)
print 'size_json'
print size_json
return Response({'closets': json, 'defaults':json_defaults,'items':item_json,'sizes':size_json,'message':'success'})
def getResponseNoId(self,request):
closets = Closet.objects.filter(owner=request.user)[0]
serializer1 = ClosetSerializer(closets, many=False)
json_defaults = JSONRenderer().render(serializer1.data)
defaults = DefaultAmounts.objects.filter(current_closet=closets.id)
serializer = ClosetDefaultsSerializer(defaults,many=True)
json = JSONRenderer().render(serializer.data)
clothing_items = ClothingItem.objects.filter(current_closet=closets.id)
item_serializer = ItemSerializer(clothing_items,many=True)
item_json = JSONRenderer().render(item_serializer.data)
size_defaults= DefaultSizes.objects.filter(current_closet=closets.id)
if len(size_defaults) == 0:
default_size = DefaultSizes(current_closet = closets)
default_size .save()
size_defaults= DefaultSizes.objects.filter(current_closet=closets.id)
size_serializer = ItemSizeSerializer(size_defaults,many=True)
size_json = JSONRenderer().render(size_serializer.data)
print 'size_json'
print size_json
return Response({'closets': json, 'defaults':json_defaults,'items':item_json,'sizes':size_json,'message':'success'})
def post(self,request,format=None):
item_name = request.data['name']
item_type = request.data['type']
item_size = request.data['size']
closet_id = request.data['closet_id']
photoDataString = request.data['photoData']
image_output = cStringIO.StringIO()
image_output.write(photoDataString.decode('base64'))
image_output.read()
image_output.seek(0) # Write decoded image to buffer
current_closet = Closet.objects.get(id=closet_id)
x = ClothingItem(name=item_name,clothing_type = item_type,size=item_size,current_closet=current_closet)
file_name = str(x.id) + '.png'
image_data = b64decode(photoDataString)
uploadedImage = ContentFile(image_data,file_name)
print uploadedImage
x.save()
x.item_image = uploadedImage
x.save()
print x.id
return Response({'failure': 'no id' }) | [
"pjryan@my.okcu.edu"
] | pjryan@my.okcu.edu |
f4b58e180e2f6563fbc03f28f8137aab6bf58d1a | f9d9afad08272fc077f9809eedb8e920fc0fe883 | /Observation/Task.py | a53536d6b3a8175d44f7420b06921eebde12790a | [] | no_license | firekg/is-it-optimal | 1c640a350d25057cf7b32602467d063e5ea65799 | 202b95ae370db7170a12ee942d6e1f0be29cca43 | refs/heads/master | 2020-05-01T13:31:35.441495 | 2019-10-27T15:34:49 | 2019-10-27T15:34:49 | 177,493,991 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,205 | py | import numpy
import Observe
import Init
import copy
import Teach
import Learn
# Eq. 6a), 6b)
def Knowledgeability_Task(hypo, feature, label, p_teacher_xy_h, p_teacher_x_h, p_y_xh, delta_g_h, phx, num_iteration):
p_learner_h_xy = Learn.Init_step(hypo, feature, label, p_y_xh, phx)
for loop in range(num_iteration):
# Calculate teacher's table
Teach.K_PTeacher_xh(hypo, feature, label, p_teacher_xy_h, p_teacher_x_h, p_learner_h_xy, delta_g_h)
# Calculate learner's table
Learn.K_PLearner_h_xy(hypo, feature, label, p_y_xh, p_learner_h_xy, p_teacher_x_h, phx)
return p_learner_h_xy
# hypo_map: The map of the hypothesis
# return: a map from hypothesis to observation * probability
def Probability_Task(hypo_table, number_hypo, number_feature, number_label, p_teacher_x_h, knowledgeability, iter=100):
feature_set = []
# New knowledgeability table
# Axis 1: index of observations
# Axis 2~3: the delta knowledegeability table
new_knowledgeability_delta_table = numpy.zeros((number_feature + 1, number_hypo, number_hypo), dtype=float)
# Assume there is a true hypo = hypo
# Get all posible hypothesis in the hypo map
for hypo_idx in range(len(hypo_table)):
# Get the observable feature set
for f in range(number_feature):
feature_set.append(f)
obs = 0
# Set the environment
num_hypo, num_feature, num_label, p_teacher_x_h, p_teacher_xy_h, p_learner_h_xy, p_y_xh, delta_g_h, phx = Init.Set(hypo_table, knowledgeability=knowledgeability)
while True:
for h in range(number_hypo):
new_knowledgeability_delta_table[obs][hypo_idx][h] = phx[h]
# Get the PT
p_learner_h_xy = Knowledgeability_Task(num_hypo, num_feature, num_label, p_teacher_xy_h, p_teacher_x_h, p_y_xh, delta_g_h, phx, iter)
# Choose a feature
feature = Observe.Get_Feature(feature_set, hypo_idx, p_teacher_x_h)
obs += 1
prob_find, true_label = Observe.Observe(hypo_table, hypo_idx, feature, p_learner_h_xy)
# Assign the p_learner_h_xy to phx
for h in range(number_hypo):
phx[h] = p_learner_h_xy[h][feature][true_label]
# remove the feature in the feature set,
# make the same feature only be observed once
feature_set.remove(feature)
if (len(feature_set) == 0):
for h in range(number_hypo):
new_knowledgeability_delta_table[obs][hypo_idx][h] = phx[h]
break
return new_knowledgeability_delta_table
def Average_Hypo(prob_map, number_hypos, number_observations):
y = []
for obs in range(number_observations):
sum = 0
for hypo_index in prob_map:
sum += prob_map[hypo_index][obs]
y.append(sum / number_hypos)
return y
| [
"noreply@github.com"
] | firekg.noreply@github.com |
cc2cf6e6ef6fb653f7cb8de7cc061e8c3a300512 | 34192b4a3964c6dec6ff34bddc06c3690c02bf98 | /Social/feed/models.py | 23c6456108e74c85bdff44c8203513abb97e3ed5 | [] | no_license | Aditya-23/The-Social | 6806882f8688f577047d79d0a02b3092cf3c4a4b | aea48c2ef816196839cad75f9dced50fa85af6f2 | refs/heads/master | 2022-10-22T11:33:22.849389 | 2019-08-11T19:31:07 | 2019-08-11T19:31:07 | 201,806,475 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,468 | py | from django.db import models
from django.db.models import Q
# Create your models here.
class Question_query(models.query.QuerySet):
def search(self, query = None):
return self.filter(
Q(question__icontains = query)
)
class Search(models.Manager):
def get_queryset(self, query = None):
return Question_query(self.model, using = self._db)
def search(self, query = None):
return self.get_queryset().search(query)
class user(models.Model):
username = models.CharField(max_length = 20, blank = False)
password = models.CharField(max_length = 20, blank = False)
firstname = models.CharField(max_length = 40, blank = False)
lastname = models.CharField(max_length = 40, blank = False)
email = models.EmailField(max_length = 40, blank = False)
#search_user = Search()
def __str__(self):
return self.username
class Question(models.Model):
question = models.CharField(max_length = 100, blank = False)
asked_user = models.ForeignKey('user', on_delete = models.CASCADE)
datetime = models.DateTimeField(auto_now = True)
objects = models.Manager()
search_question = Search()
def __str__(self):
return self.question
class Answer(models.Model):
answer = models.CharField(max_length = 1000, blank = False)
datetime = models.DateTimeField(auto_now = True)
question = models.ForeignKey('Question', on_delete = models.CASCADE)
answered_user = models.ForeignKey('user', on_delete = models.CASCADE)
def __str__(self):
return self.answer | [
"adityamysore002@gmail.com"
] | adityamysore002@gmail.com |
9c17bdc0d3beabedf1313533658ffff019329cce | 8f5d2fb45d6452fc6df00b12fa0bd45446d1029b | /lessons/models.py | 1227de579706645ec9701f1a78da6f80fcc40145 | [] | no_license | luminyanko/renshuu | 4bcbd0f10e0a4a175ab92783e5346ac4b1c44927 | 54a98f59a7d2971e3ba20d5406b3e7ad8c482408 | refs/heads/master | 2023-06-01T20:40:30.090022 | 2021-06-21T08:47:19 | 2021-06-21T08:47:19 | 377,913,945 | 0 | 0 | null | 2021-06-21T08:44:52 | 2021-06-17T17:39:11 | Python | UTF-8 | Python | false | false | 734 | py | from django.db import models
from django.urls import reverse
from django.utils import timezone
from django.contrib.auth.models import User
# Create your models here.
class Tag(models.Model):
tag_name = models.CharField(max_length=50)
def __str__(self):
return self.tag_name
class Lesson(models.Model):
title = models.CharField(max_length=150)
content = models.TextField()
date_created = models.DateTimeField(default=timezone.now)
author = models.ForeignKey(User, on_delete=models.CASCADE)
tag = models.ForeignKey(Tag, on_delete=models.PROTECT)
def get_absolute_url(self):
return reverse('lesson-detail', kwargs={'pk': self.pk})
def __str__(self):
return self.title
| [
"luminyanko@gmail.com"
] | luminyanko@gmail.com |
1c125d93dacd44efb23a7c8db2107c3a5838f0bb | 25ac403b85c141bde644f92785d086b2049ccbff | /数据/utf8csv2excel.py | 5f12f35818e9db1abf5c1b20b9a128038025af4c | [] | no_license | void126rlz/SHUJRA | 62ac47313a2c42dc439cb027a73376bfbe116ec4 | b7451af92df4bfdd80e71a2e7cbab7eecfb28f23 | refs/heads/main | 2023-07-17T06:26:01.119303 | 2021-09-07T13:39:25 | 2021-09-07T13:39:25 | 379,532,808 | 1 | 1 | null | 2021-07-11T08:42:34 | 2021-06-23T08:28:06 | Jupyter Notebook | UTF-8 | Python | false | false | 1,797 | py | import os # sys
import pandas as pd
home = r'D:\tfp\project\新致软件\3岗位推荐算法\from新致\数据'
fileNamesCSV = os.listdir(home + r'\历史18天的数据') # \*.csv
print(fileNamesCSV)
# !dir D:\tfp\project\新致软件\3岗位推荐算法\from新致\数据\utf8
fileNames = [s[0:-4] for s in fileNamesCSV]
print(fileNames)
# InvalidWorksheetName: Excel worksheet name 'ttyc_personel_educational_experience' must be <= 31 chars.
# print(len('ttyc_personel_educational_experience'))
# [len(f) for f in fileNamesCSV] # <34-3=31
# [18, 14, 18, 41, 24, 37, 34, 17, 23]
[len(f) for f in fileNames]
fileNames=[
# 'ttyc_candidate',
# 'ttyc_label',
# 'ttyc_personnel',
'ttyc_personnel_educational_experience',
'ttyc_personnel_label',
'ttyc_personnel_project_experience',
'ttyc_personnel_work_experience',
'ttyc_position',
'ttyc_position_label'
]
for f in fileNames:
print(f)
df= pd.read_csv(home + '\\历史18天的数据\\' + f + '.csv', error_bad_lines=False )
df.to_excel(home + '\\excel18\\' + f + '.xlsx', sheet_name=f[5:36],index=False)
# f = 'ttyc_personnel'
df= pd.read_csv(home + '\\历史18天的数据\\ttyc_personnel-tfp.csv', error_bad_lines=False )
df.to_excel(home + '\\excel18\\ttyc_personnel-tfp.xlsx', sheet_name='ttyc_personnel',index=False)
# ;号分隔的cvs文件
fileNames2 = ['ttyc_position', 'ttyc_position_label']
for f in fileNames2:
print(f)
df= pd.read_csv(home + '\\utf8\\' + f + '.csv', sep=';') # quotechar='"',
df.to_excel(home + '\\excel\\' + f + '.xlsx', sheet_name=f,index=False)
f = 'ttyc_personnel_project_experienceT'
df= pd.read_csv(home + '\\utf8\\' + f + '.csv')
df.to_excel(home + '\\excel\\' + f + '.xlsx', sheet_name=f[5:36],index=False)
| [
"noreply@github.com"
] | void126rlz.noreply@github.com |
7a1df63cd632b5b6f4ccaeaeee6eff6164e582d7 | bffcfa6103ee72d7ac394c14aa861e60616c7ab8 | /pytorch3d/datasets/__init__.py | 1687213018a29e5d75a4c5490368d52e5f4d893a | [
"BSD-3-Clause"
] | permissive | Amit2016-17/pytorch3d | ccac686bc1a3caeb4bd0f38519fbcb83f816501d | 7944d24d4872bdb01b821450840049e28d0ce12b | refs/heads/master | 2022-11-25T10:40:14.409087 | 2020-08-05T13:58:53 | 2020-08-05T14:00:20 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 255 | py | # Copyright (c) Facebook, Inc. and its affiliates. All rights reserved.
from .r2n2 import R2N2, BlenderCamera
from .shapenet import ShapeNetCore
from .utils import collate_batched_meshes
__all__ = [k for k in globals().keys() if not k.startswith("_")]
| [
"facebook-github-bot@users.noreply.github.com"
] | facebook-github-bot@users.noreply.github.com |
7dfcead14cfcc41518ec35eaa9c96ca9cfbc0be3 | 8fb846f4f4ac5fd417489d731eae8a8a1bdc77c3 | /rllab/misc/console.py | b32d21a249a3d389e0aef97f641591cdb13bb35a | [
"LicenseRef-scancode-generic-cla",
"MIT"
] | permissive | zhongwen/rllab | 0a9f9ea2d8995037b83aaae5853a299d5cf9e432 | d8239c05179fcc55d865db7ce933defa3baae24d | refs/heads/master | 2021-01-14T08:36:37.272071 | 2016-08-17T12:29:00 | 2016-08-17T12:29:00 | 65,801,245 | 1 | 1 | null | 2016-08-16T08:18:47 | 2016-08-16T08:18:46 | null | UTF-8 | Python | false | false | 5,514 | py | import sys
import time
import os
import errno
import shlex
import pydoc
import inspect
color2num = dict(
gray=30,
red=31,
green=32,
yellow=33,
blue=34,
magenta=35,
cyan=36,
white=37,
crimson=38
)
def colorize(string, color, bold=False, highlight=False):
attr = []
num = color2num[color]
if highlight:
num += 10
attr.append(str(num))
if bold:
attr.append('1')
return '\x1b[%sm%s\x1b[0m' % (';'.join(attr), string)
def mkdir_p(path):
try:
os.makedirs(path)
except OSError as exc: # Python >2.5
if exc.errno == errno.EEXIST and os.path.isdir(path):
pass
else:
raise
def log(s): # , send_telegram=False):
print s
sys.stdout.flush()
class SimpleMessage(object):
def __init__(self, msg, logger=log):
self.msg = msg
self.logger = logger
def __enter__(self):
print self.msg
self.tstart = time.time()
def __exit__(self, etype, *args):
maybe_exc = "" if etype is None else " (with exception)"
self.logger("done%s in %.3f seconds" %
(maybe_exc, time.time() - self.tstart))
MESSAGE_DEPTH = 0
class Message(object):
def __init__(self, msg):
self.msg = msg
def __enter__(self):
global MESSAGE_DEPTH # pylint: disable=W0603
print colorize('\t' * MESSAGE_DEPTH + '=: ' + self.msg, 'magenta')
self.tstart = time.time()
MESSAGE_DEPTH += 1
def __exit__(self, etype, *args):
global MESSAGE_DEPTH # pylint: disable=W0603
MESSAGE_DEPTH -= 1
maybe_exc = "" if etype is None else " (with exception)"
print colorize('\t' * MESSAGE_DEPTH + "done%s in %.3f seconds" % (maybe_exc, time.time() - self.tstart), 'magenta')
def prefix_log(prefix, logger=log):
return lambda s: logger(prefix + s)
def tee_log(file_name):
f = open(file_name, 'w+')
def logger(s):
log(s)
f.write(s)
f.write('\n')
f.flush()
return logger
def collect_args():
splitted = shlex.split(' '.join(sys.argv[1:]))
return {arg_name[2:]: arg_val
for arg_name, arg_val in zip(splitted[::2], splitted[1::2])}
def type_hint(arg_name, arg_type):
def wrap(f):
meta = getattr(f, '__tweak_type_hint_meta__', None)
if meta is None:
f.__tweak_type_hint_meta__ = meta = {}
meta[arg_name] = arg_type
return f
return wrap
def tweak(fun_or_val, identifier=None):
if callable(fun_or_val):
return tweakfun(fun_or_val, identifier)
return tweakval(fun_or_val, identifier)
def tweakval(val, identifier):
if not identifier:
raise ValueError('Must provide an identifier for tweakval to work')
args = collect_args()
for k, v in args.iteritems():
stripped = k.replace('-', '_')
if stripped == identifier:
log('replacing %s in %s with %s' % (stripped, str(val), str(v)))
return type(val)(v)
return val
def tweakfun(fun, alt=None):
"""Make the arguments (or the function itself) tweakable from command line.
See tests/test_misc_console.py for examples.
NOTE: this only works for the initial launched process, since other processes
will get different argv. What this means is that tweak() calls wrapped in a function
to be invoked in a child process might not behave properly.
"""
cls = getattr(fun, 'im_class', None)
method_name = fun.__name__
if alt:
cmd_prefix = alt
elif cls:
cmd_prefix = cls + '.' + method_name
else:
cmd_prefix = method_name
cmd_prefix = cmd_prefix.lower()
args = collect_args()
if cmd_prefix in args:
fun = pydoc.locate(args[cmd_prefix])
if type(fun) == type:
argspec = inspect.getargspec(fun.__init__)
else:
argspec = inspect.getargspec(fun)
# TODO handle list arguments
defaults = dict(
zip(argspec.args[-len(argspec.defaults or []):], argspec.defaults or []))
replaced_kwargs = {}
cmd_prefix += '-'
if type(fun) == type:
meta = getattr(fun.__init__, '__tweak_type_hint_meta__', {})
else:
meta = getattr(fun, '__tweak_type_hint_meta__', {})
for k, v in args.iteritems():
if k.startswith(cmd_prefix):
stripped = k[len(cmd_prefix):].replace('-', '_')
if stripped in meta:
log('replacing %s in %s with %s' % (stripped, str(fun), str(v)))
replaced_kwargs[stripped] = meta[stripped](v)
elif stripped not in argspec.args:
raise ValueError(
'%s is not an explicit parameter of %s' % (stripped, str(fun)))
elif stripped not in defaults:
raise ValueError(
'%s does not have a default value in method %s' % (stripped, str(fun)))
elif defaults[stripped] is None:
raise ValueError(
'Cannot infer type of %s in method %s from None value' % (stripped, str(fun)))
else:
log('replacing %s in %s with %s' % (stripped, str(fun), str(v)))
# TODO more proper conversions
replaced_kwargs[stripped] = type(defaults[stripped])(v)
def tweaked(*args, **kwargs):
all_kw = dict(zip(argspec[0], args) +
kwargs.items() + replaced_kwargs.items())
return fun(**all_kw)
return tweaked
| [
"dementrock@gmail.com"
] | dementrock@gmail.com |
ebbf514c6ec2b30cad1b8c147dfef1377c951533 | 82e5c56806e75a51097a51cb721df1ef07376e6f | /initial_freia/fun.py | 6b89a7f268c642c02b65a655544d8257951ff747 | [] | no_license | hahahasan/MSc-pre-process | 77a2fcdf795830d2b37b2d538e3638d29bd55e80 | ba69d58bc1f9dd7becea431ce9a328b2843d8b66 | refs/heads/master | 2020-03-21T05:43:07.501035 | 2018-06-21T14:14:30 | 2018-06-21T14:14:30 | 138,175,816 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 697 | py | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Wed Jun 13 11:06:10 2018
@author: hm1234
"""
import numpy as np
import matplotlib.pyplot as plt
from mpl_toolkits.mplot3d import Axes3D
rng = np.linspace(0.01, np.sqrt(2), 500)
rng2 = np.arange(10,500)
max_rng = np.amax(rng2)
hi = []
for j in rng2:
n = j
tmp = []
for i in rng:
a = i
b = '**a'
c = str(a) + n*b
tmp.append(eval(c))
if j % 23 == 0:
print('{0:1.0f}'.format(j/max_rng *100), '%')
hi.append(tmp)
print('Done!')
X_rng, Y_rng = np.meshgrid(rng2, rng)
fig = plt.figure()
ax = fig.add_subplot(111, projection='3d')
ax.plot_surface(X_rng, Y_rng, np.array(hi).T)
| [
"hm1234@york.ac.uk"
] | hm1234@york.ac.uk |
9f800cece6947820a5e6c1324c4b395ea3efdf40 | a1c397a8b4dccfbef8b91a67d6910a58ff6aa98e | /02-19-Cuma/tryExcept.py | 13bb01e466c735f27fe0a613acc049813ea2ed45 | [
"MIT"
] | permissive | hhsalik/staj | f0c9a17064c444fa084a102e4c36050ff62fdb4e | d0ee95d5e77a7d7a1f16611d49c87be429a25b31 | refs/heads/master | 2023-04-29T15:17:17.928023 | 2021-05-10T19:25:35 | 2021-05-10T19:25:35 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 180 | py | try:
answer = 10 / 0
number = int(input("Enter a number: "))
print(number)
except ZeroDivisionError as err:
print(err)
except ValueError:
print("invalid input") | [
"cihat02020202@gmail.com"
] | cihat02020202@gmail.com |
4b6acdbc898c6ac10a32ac3ebabe2a347cac037f | a5846332d42c8705d054792bdac219560d77d32f | /Test/test.py | 5dfe4d8d72022058c8dfbfa854f2b08bdaab2bc1 | [] | no_license | Jinsung-Jeon/Deep_Learning_Code | 2927a65adc9c7ef125603d186d4d5f38cebcfade | a87658c4beb823f9c34019e28cad736b259a7747 | refs/heads/master | 2020-12-13T02:35:52.306209 | 2020-03-06T07:01:11 | 2020-03-06T07:01:11 | 234,288,214 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 23,159 | py | # -*- coding: utf-8 -*-
"""
Created on Thu Jan 23 14:32:43 2020
@author: Jinsung
"""
#Chap1 test
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap1-Regression analysis Estimation of the number of rings in abalone/abalone.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap1-Regression analysis Estimation of the number of rings in abalone')
abalone_exec()
#Chap2 test
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star/pulsar.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star')
pulsar_exec()
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star/pulsar_ext.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star')
pulsar_exec(adjust_ratio=True)
#Chap3 test
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap3-Multi Classification/steel_test.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap3-Multi Classification')
steel_exec()
#Chap4 test
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure/mlp.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure')
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap1-Regression analysis Estimation of the number of rings in abalone/abalone.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap1-Regression analysis Estimation of the number of rings in abalone')
set_hidden([])
abalone_exec(epoch_count=50, report=10)
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure/mlp.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure')
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star/pulsar.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star')
set_hidden(6)
pulsar_exec(epoch_count=50, report=10)
set_hidden([12,6])
pulsar_exec(epoch_count=50, report=10)
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure/mlp.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure')
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star/pulsar_ext.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap2-Binary Classification predicting a pulsar star')
set_hidden([12,6])
pulsar_exec(epoch_count=50, report=10, adjust_ratio=True)
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure/mlp.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap4-MLP based structure')
runfile('C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap3-Multi Classification/steel_test.py',
wdir='C:/Users/Jinsung/Documents/Deep_Learning_Code/Chap3-Multi Classification')
LEARNING_RATE = 0.0001
set_hidden([12,6,4])
steel_exec(epoch_count=50, report=10)
#Chap5 test
ad = AbaloneDataset()
am = MlpModel('abalone_model',ad,[])
am.exec_all(epoch_count=10, report=2)
pd = PulsarDataset()
pm = MlpModel('pulsar_model', pd, [4])
pm.exec_all()
pm.visualize(5)
sd = SteelDataset()
sm = MlpModel('steel_model', sd, [12,7])
sm.exec_all(epoch_count=50, report=10)
psd = PulsarSelectDataset()
psm = MlpModel('pulsar_select_model', psd, [4])
psm.exec_all()
fd = FlowersDataset()
fm = MlpModel('flowers_model_1', fd, [10]) #같은 추정확률분포를 통해 언제나 민들레라는 답을 냈다.
fm.exec_all(epoch_count=10, report=2)
fm2 = MlpModel('flowers_model_2', fd, [30,10])
fm2.exec_all(epoch_count=10, report=2)
#Chap6 test
od = Office31Dataset()
om1 = MlpModel('office31_model_1', od, [10])
om1.exec_all(epoch_count=20, report=10)
om2 = MlpModel('office31_model_2', od, [64,32,10])
om2.exec_all(epoch_count=20, report=10, learning_rate=0.0001)
om3 = MlpModel('office31_model_3', od, [64,32,10])
om3.use_adam = True
om3.exec_all(epoch_count=50, report=10, learning_rate=0.0001)
#Chap7 test
fd = FlowersDataset([96, 96], [96, 96, 3])
od = Office31Dataset([96, 96], [96, 96, 3])
fm1 = CnnBasicModel('flowers_model_1', fd, [30, 10])
fm1.exec_all(epoch_count=10, report=2)
fm2 = CnnBasicModel('flowers_model_2', fd, [['full', {'width':30}],['full', {'width':10}]])
fm2.use_adam=False
fm2.exec_all(epoch_count = 10, report = 2)
fm3 = CnnBasicModel('flowers_model_3', fd, [['conv', {'ksize':5, 'chn':6}],
['max', {'stride':4}],
['conv', {'ksize':3, 'chn':12}],
['avg', {'stride':2}]],
True)
fm3.exec_all(epoch_count = 10, report=2)
om1 = CnnBasicModel('officie31_model_1', od,
[['conv', {'ksize':3, 'chn':6}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':12}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':24}],
['avg', {'stride':3}]])
om1.exec_all(epoch_count=10, report =2)
om2 = CnnBasicModel('officie31_model_2', od,
[['conv', {'ksize':3, 'chn':6, 'actfunc':'sigmoid'}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':12, 'actfunc':'sigmoid'}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':24, 'actfunc':'sigmoid'}],
['avg', {'stride':3}]])
om2.exec_all(epoch_count=10, report =2)
om3 = CnnBasicModel('officie31_model_3', od,
[['conv', {'ksize':3, 'chn':6, 'actfunc':'tanh'}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':12, 'actfunc':'tanh'}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':24, 'actfunc':'tanh'}],
['avg', {'stride':3}]])
om3.exec_all(epoch_count=10, report =2)
#Chap8 test
fd = FlowersDataset([96, 96], [96, 96, 3])
od = Office31Dataset([96, 96], [96, 96, 3])
fm1 = CnnRegModel('flowers_model_1', fd, [30, 10])
fm1.exec_all(epoch_count=10, report=2, show_params=True)
fm2 = CnnRegModel('flowers_model_2', fd, [30, 10], l2_decay=0.1)
fm2.exec_all(epoch_count=10, show_cnt=0, show_params=True)
fm3 = CnnRegModel('flowers_model_3', fd, [30, 10], l1_decay=0.1)
fm3.exec_all(epoch_count=10, show_cnt=0, show_params=True)
cnn1 = [['conv', {'ksize':3, 'chn':6}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':12}],
['max', {'stride':2}],
['conv', {'ksize':3, 'chn':24}],
['avg', {'stride':3}]]
fcnn1 = CnnRegModel('flowers_cnn_1')
fcnn1.exec_all(epoch_count=10, report=2)
cnn2 = [['conv', {'ksize':3, 'chn':6}],
['max', {'stride':2}],
['dropout', {'keep_prob':0.6}],
['conv', {'ksize':3, 'chn':12}],
['max', {'stride':2}],
['dropout', {'keep_prob':0.6}],
['conv', {'ksize':3, 'chn':24}],
['avg', {'stride':3}],
['dropout', {'keep_prob':0.6}]]
fcnn2 = CnnRegModel('flowers_cnn_2',fd, cnn2)
fcnn2.exec_all(epoch_count=10, report=2, show_cnt=0)
cnn3 = [['noise', {'type':'normal','mean':0,'std':0.01}],
['conv', {'ksize':3, 'chn':6}],
['max', {'stride':2}],
['noise', {'type':'normal','mean':0,'std':0.01}],
['conv', {'ksize':3, 'chn':12}],
['max', {'stride':2}],
['noise', {'type':'normal','mean':0,'std':0.01}],
['conv', {'ksize':3, 'chn':24}],
['avg', {'stride':3}]]
fcnn3 = CnnRegModel('flowers_cnn_3',fd, cnn3)
fcnn3.exec_all(epoch_count=10, report=2, show_cnt=0)
cnn4 = [['batch_normal'],
['conv', {'ksize':3, 'chn':6}],
['max', {'stride':2}],
['batch_normal'],
['conv', {'ksize':3, 'chn':12}],
['max', {'stride':2}],
['batch_normal'],
['conv', {'ksize':3, 'chn':24}],
['avg', {'stride':3}]]
fcnn4 = CnnRegModel('flowers_cnn_4',fd, cnn4)
fcnn4.exec_all(epoch_count=10, report=2, show_cnt=0)
od = Office31Dataset([96, 96], [96, 96, 3])
ocnn1 = CnnRegModel('office31_cnn_1', od, cnn1)
ocnn2 = CnnRegModel('office31_cnn_2', od, cnn1)
ocnn3 = CnnRegModel('office31_cnn_3', od, cnn1)
ocnn4 = CnnRegModel('office31_cnn_4', od, cnn1)
ocnn1.exec_all(epoch_count=10, show_cnt=0)
ocnn2.exec_all(epoch_count=10, show_cnt=0)
ocnn3.exec_all(epoch_count=10, show_cnt=0)
ocnn4.exec_all(epoch_count=10, show_cnt=0)
# Chap9
# inception-v3 model
imagenet = DummyDataset('imagenet', 'select', [299,299,3], 200)
CnnExtModel.set_macro('v3_preproc',
['serial',
['conv', {'ksize':3, 'stride':2, 'chn':32, 'padding':'VALID'}],
['conv', {'ksize':3, 'chn':32, 'padding':'VALID'}],
['conv', {'ksize':3, 'chn':64, 'padding':'SAME'}],
['max', {'ksize':3, 'stride':2, 'padding':'VALID'}],
['conv', {'ksize':1, 'chn':80, 'padding':'VALID'}],
['max', {'ksize':3, 'stride':2, 'padding':'VALID'}]])
CnnExtModel.set_macro('v3_inception1',
['parallel',
['conv', {'ksize':1, 'chn':64}],
['serial',
['conv', {'ksize':1, 'chn':48}],
['conv', {'ksize':5, 'chn':64}]],
['serial',
['conv', {'ksize':1, 'chn':64}],
['conv', {'ksize':3, 'chn':96}],
['conv', {'ksize':3, 'chn':96}]],
['serial',
['avg', {'ksize':3, 'stride':1}],
['conv', {'ksize':1, 'chn':'#chn'}]]])
CnnExtModel.set_macro('v3_resize1',
['parallel',
['conv', {'ksize':1, 'stride':2, 'chn':384}],
['serial',
['conv', {'ksize':1, 'chn':64}],
['conv', {'ksize':3, 'chn':96}],
['conv', {'ksize':3, 'stride':2, 'chn':96}]],
['max', {'ksize':3, 'stride':2}]])
CnnExtModel.set_macro('v3_inception2',
['parallel',
['conv', {'ksize':1, 'chn':192}],
['serial',
['conv', {'ksize':[1,1], 'chn':'#chn'}],
['conv', {'ksize':[1,7], 'chn':'#chn'}],
['conv', {'ksize':[7,1], 'chn':192}]],
['serial',
['conv', {'ksize':[1,1], 'chn':'#chn'}],
['conv', {'ksize':[7,1], 'chn':'#chn'}],
['conv', {'ksize':[1,7], 'chn':'#chn'}],
['conv', {'ksize':[7,1], 'chn':'#chn'}],
['conv', {'ksize':[1,7], 'chn':192}]],
['serial',
['avg', {'ksize':3, 'stride':1}],
['conv', {'ksize':1, 'chn':192}]]])
CnnExtModel.set_macro('v3_resize2',
['parallel',
['serial',
['conv', {'ksize':1, 'chn':192}],
['conv', {'ksize':3, 'stride':2, 'chn':320}]],
['serial',
['conv', {'ksize':[1,1], 'chn':192}],
['conv', {'ksize':[1,7], 'chn':192}],
['conv', {'ksize':[7,1], 'chn':192}],
['conv', {'ksize':[3,3], 'stride':[2,2], 'chn':192}]],
['max', {'ksize':3, 'stride':2}]])
CnnExtModel.set_macro('v3_inception3',
['parallel',
['conv', {'ksize':1, 'chn':320}],
['serial',
['conv', {'ksize':[3,3], 'chn':384}],
['parallel',
['conv', {'ksize':[1,3], 'chn':384}],
['conv', {'ksize':[3,1], 'chn':384}]]],
['serial',
['conv', {'ksize':[1,1], 'chn':448}],
['conv', {'ksize':[3,3], 'chn':384}],
['parallel',
['conv', {'ksize':[1,3], 'chn':384}],
['conv', {'ksize':[3,1], 'chn':384}]]],
['serial',
['avg', {'ksize':3, 'stride':1}],
['conv', {'ksize':1, 'chn':192}]]])
CnnExtModel.set_macro('v3_postproc',
['serial',
['avg', {'stride':8}],
['dropout', {'keep_prob':0.7}]])
CnnExtModel.set_macro('inception_v3',
['serial',
['custom', {'name':'v3_preproc'}],
['custom', {'name':'v3_inception1', 'args':{'#chn':32}}],
['custom', {'name':'v3_inception1', 'args':{'#chn':64}}],
['custom', {'name':'v3_inception1', 'args':{'#chn':64}}],
['custom', {'name':'v3_resize1'}],
['custom', {'name':'v3_inception2', 'args':{'#chn':128}}],
['custom', {'name':'v3_inception2', 'args':{'#chn':160}}],
['custom', {'name':'v3_inception2', 'args':{'#chn':160}}],
['custom', {'name':'v3_inception2', 'args':{'#chn':192}}],
['custom', {'name':'v3_resize2'}],
['custom', {'name':'v3_inception3'}],
['custom', {'name':'v3_inception3'}],
['custom', {'name':'v3_postproc'}]])
inception_v3 = CnnExtModel('inception_v3', imagenet, [['custom', {'name':'inception_v3'}]], dump_structure=True)
fd = FlowersDataset([96, 96], [96, 96, 3])
CnnExtModel.set_macro('flower_preproc',
['serial',
['conv', {'ksize':3, 'stride':2, 'chn':6, 'actions':'#act'}]])
CnnExtModel.set_macro('flower_inception1',
['parallel',
['conv', {'ksize':1, 'chn':4, 'actions':'#act'}],
['conv', {'ksize':3, 'chn':6, 'actions':'#act'}],
['serial',
['conv', {'ksize':3, 'chn':6, 'actions':'#act'}],
['conv', {'ksize':3, 'chn':6, 'actions':'#act'}]],
['serial',
['avg', {'ksize':3, 'stride':1}],
['conv', {'ksize':1, 'chn':4, 'actions':'#act'}]]])
CnnExtModel.set_macro('flower_resize',
['parallel',
['conv', {'ksize':1, 'stride':2, 'chn':12, 'actions':'#act'}],
['serial',
['conv', {'ksize':3, 'chn':12, 'actions':'#act'}],
['conv', {'ksize':3, 'stride':2, 'chn':12, 'actions':'#act'}]],
['avg', {'ksize':3, 'stride':2}]])
CnnExtModel.set_macro('flower_inception2',
['parallel',
['conv', {'ksize':1, 'chn':8, 'action':'#act'}],
['serial',
['conv', {'ksize':[3,3], 'chn':8, 'actions':'#act'}],
['parallel',
['conv', {'ksize':[1,3], 'chn':8, 'actions':'#act'}],
['conv', {'ksize':[3,1], 'chn':8, 'actions':'#act'}]]],
['serial',
['conv', {'ksize':[1,1], 'chn':8, 'actions':'#act'}],
['conv', {'ksize':[3,3], 'chn':8, 'actions':'#act'}],
['parallel',
['conv', {'ksize':[1,3], 'chn':8, 'actions':'#act'}],
['conv', {'ksize':[3,1], 'chn':8, 'actions':'#act'}]]],
['serial',
['avg', {'ksize':3, 'stride':1}],
['conv', {'ksize':1, 'chn':8, 'actions':'#act'}]]])
CnnExtModel.set_macro('flower_postproc',
['serial',
['avg', {'stride':6}],
['dropout', {'keep_prob':0.7}]])
CnnExtModel.set_macro('inception_flower',
['serial',
['custom', {'name':'flower_preproc', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_inception1', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_resize', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_inception1', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_resize', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_inception2', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_resize', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_inception2', 'args':{'#act':'#act'}}],
['custom', {'name':'flower_postproc', 'args':{'#act':'#act'}}]])
conf_flower_LA = ['custom', {'name':'inception_flower', 'args':{'#act':'LA'}}]
model_flower_LA = CnnExtModel('model_flower_LA', fd, conf_flower_LA, dump_structure=True)
model_flower_LA.exec_all(report=2)
conf_flower_LAB = ['custom', {'name':'inception_flower', 'args':{'#act':'LAB'}}]
model_flower_LAB = CnnExtModel('model_flower_LAB', fd, conf_flower_LAB, dump_structure=False)
model_flower_LAB.exec_all(epoch_count=10, report=2)
#Chap10
ad = AutomataDataset()
am_4 = RnnBasicModel('am_4', ad, ['rnn', {'recur_size':4, 'outseq':False}])
am_16 = RnnBasicModel('am_16', ad, ['rnn', {'recur_size':16, 'outseq':False}])
am_64 = RnnBasicModel('am_64', ad, ['rnn', {'recur_size':64, 'outseq':False}])
am_4.exec_all(epoch_count=10, report=2)
am_16.exec_all(epoch_count=10, report=2)
am_64.exec_all(epoch_count=10, report=2)
am_64_drop = RnnBasicModel('am_64_drop', ad, [['rnn', {'recur_size':64, 'outseq':False}],['dropout', {'keep_prob':0.5}]])
am_64_drop.exec_all(epch_count=10, report=2)
#Chap11
ad = AutomataDataset()
am_4 = RnnLstmModel('am_4', ad, ['lstm', {'recur_size':64, 'outseq':False}])
am_4.exec_all(epoch_count=10, report=2)
usd_10_10 = UrbanSoundDataset(10, 10)
usd_10_100 = UrbanSoundDataset(10, 100)
conf_basic = ['rnn', {'recur_size':20, 'outseq':False}]
conf_lstm = ['lstm', {'recur_size':20, 'outseq':False}]
conf_state = ['lstm', {'recur_size':20, 'outseq':False, 'use_state':True}]
us_basic_10_10 = RnnLstmModel('us_basic_10_10', usd_10_10, conf_basic)
us_lstm_10_10 = RnnLstmModel('us_lstm_10_10', usd_10_10, conf_lstm)
us_state_10_10 = RnnLstmModel('us_state_10_10', usd_10_10, conf_state)
us_basic_10_100 = RnnLstmModel('us_basic_10_100', usd_10_100, conf_basic)
us_lstm_10_100 = RnnLstmModel('us_lstm_10_100', usd_10_100, conf_lstm)
us_state_10_100 = RnnLstmModel('us_state_10_100', usd_10_100, conf_state)
us_basic_10_10.exec_all(epoch_count=10, report=2)
us_lstm_10_10.exec_all(epoch_count=10, report=2)
us_state_10_10.exec_all(epoch_count=10, report=2, show_cnt=0)
#Chap12
vsd = np.load('C:\\Users\\Jinsung\\Documents\\Deep_Learning_Code\\Datasets\\chap12\\cache\\AstarIsBorn1937.mp4.npy')
conf1 = [['seqwrap', ['avg', {'stride':30}],
['conv', {'ksize':3, 'chn':12}],
['full', {'width':16}]],
['lstm', {'recur_size':8}]]
vsm1 = RnnExtModel('vsm1', vsd, conf1)
vsm1.exec_all(epoch_count=10, report=2, show_cnt=3)
vsd.shape
#Chap13
mset_all = MnistAutoDataset(train_ratio=1.00)
mset_1p = MnistAutoDataset(train_ratio=0.01)
conf_mlp = [['full',{'width':10}]]
mnist_mlp_all = RnnExtModel('mnist_mlp_all', mset_all, conf_mlp)
mnist_mlp_all.exec_all(epoch_count=10, report=2)
conf_auto = {
'encoder': [['full', {'width':10}]],
'decoder': [['full', {'width':784}]],
'supervised': [['full', {'width':10}]]
}
mnist_auto_1 = Autoencoder('mnist_auto_1',mset_1p, conf_auto)
mnist_auto_1.autoencode(epoch_count=10, report=2)
mnist_auto_1.exec_all(epoch_count=10, report=2)
mnist_auto_fix = Autoencoder('mnist_auto_fix', mset_1p, conf_auto, fix_encoder=True)
mnist_auto_fix.autoencode(epoch_count=10, report=5)
mnist_auto_fix.exec_all(epoch_count=10, report=5)
conf_auto_2 = {
'encoder': [['full', {'width':64}], ['full', {'width':10}]],
'decoder': [['full', {'width':64}], ['full', {'width':784}]],
'supervised': [['full', {'width':10}]]
}
mnist_auto_2 = Autoencoder('mnist_auto_2',mset_1p, conf_auto_2)
mnist_auto_2.autoencode(epoch_count=10, report=2)
mnist_auto_2.exec_all(epoch_count=10, report=2)
conf_hash_1 = {
'encoder': [['full', {'width':10, 'actfunc':'sigmoid'}]],
'decoder': [['full', {'width':784}]],
'supervised': []
}
mnist_hash_1 = Autoencoder('mnist_hash_1',mset_1p, conf_hash_1)
mnist_hash_1.autoencode(epoch_count=10, report=2)
mnist_hash_1.semantic_hashing_index()
mnist_hash_1.semantic_hashing_search()
conf_hash_2 = {
'encoder': [['full', {'width':64}],['full', {'width':10, 'actfunc':'sigmoid'}]],
'decoder': [['full', {'width':64}],['full', {'width':784}]],
'supervised': []
}
mnist_hash_2 = Autoencoder('mnist_hash_2',mset_1p, conf_hash_2)
mnist_hash_2.autoencode(epoch_count=10, report=2)
mnist_hash_2.semantic_hashing_index()
mnist_hash_2.semantic_hashing_search()
mnist_hash_2.autoencode(epoch_count=40, report=10)
mnist_hash_2.semantic_hashing_index()
mnist_hash_2.semantic_hashing_search()
#Chap14
mnist_eng = MnistEngDataset()
conf_eng1 = {
'encoder': [['full', {'width':10}]],
'decoder': [['lstm', {'recur_size':32, 'inseq':False,
'outseq':True, 'timesteps':6}],
['seqwrap', ['full', {'width':27, 'actfunc':'none'}]]]
}
encdec_eng1 = EncoderDecoder('encdec_eng1', mnist_eng, conf_eng1)
encdec_eng1.exec_1_step(epoch_count=10, report=2)
conf_eng2 = {
'encoder': [['full', {'width':10}],
['batch_normal'],
['full', {'width':10}]],
'decoder': [['lstm', {'recur_size':32, 'inseq':False,
'outseq':True, 'timesteps':6}],
['seqwrap', ['full', {'width':27, 'actfunc':'none'}]]]
}
encdec_eng2 = EncoderDecoder('encdec_eng2', mnist_eng, conf_eng2)
encdec_eng2.exec_1_step(epoch_count=10, report=2)
encdec_eng2_2 = EncoderDecoder('encdec_eng2_2', mnist_eng, conf_eng2)
encdec_eng2_2.exec_2_step(epoch_count=10, report=5)
encdec_eng2_3 = EncoderDecoder('encdec_eng2_3', mnist_eng, conf_eng2)
encdec_eng2_3.exec_3_step(epoch_count=10, report=5)
#Chap15
dset_pic_gogh = GanDatasetPicture('gogh.jpg')
dset_pic_jungsun = GanDatasetPicture('jungsun.jpg')
print(dset_pic_gogh)
print(dset_pic_jungsun)
conf_pic = {
'seed_shape': [16],
'generator': [['full', {'width':64}],
['full', {'width':32*32*3, 'actfunc':'sigmoid'}]],
'discriminor': [['full', {'width':64}],
['full', {'width':1, 'actfunc':'none'}]]
}
gan_pic_gogh = Gan("gan_pic_gogh", dset_pic_gogh, conf_pic, dump_structure=True)
gan_pic_gogh.exec_all(epoch_count=100, report=20)
| [
"jjsjjs0902@naver.com"
] | jjsjjs0902@naver.com |
da731a7f36e0c4abfe56e401d84517f945860e52 | 14d5bccb090070fc212651d017b71a1f2c6fadce | /mysite/myapi/models.py | d29e391516d08b76c141036c24658c027c6e2666 | [] | no_license | ryyvntong/SuperheroAPI | 223c78abd0a46a645ce36a871a7df5008dac6d02 | 6dd2e63a44bf7fc251abacd90cf3782fafb24e9d | refs/heads/master | 2020-12-31T21:32:41.788871 | 2020-02-07T22:02:36 | 2020-02-07T22:02:36 | 239,028,186 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 221 | py | from django.db import models
# Create your models here.
class Hero(models.Model):
name = models.CharField(max_length=60)
alias = models.CharField(max_length=60)
def __str__(self):
return self.name
| [
"56896048+ryyvntong@users.noreply.github.com"
] | 56896048+ryyvntong@users.noreply.github.com |
80aa265104445e40404828f18f155d2af0bd25b2 | c2285444392e8eb1904255b82f62b49f317aca07 | /scripts/load_tags.py | 59a372e1337682c26b6ce0ab3ed0e6c41f62fb28 | [] | no_license | tlskr/tagger | ced0ed36437bb29fe488eb2fae8b03314c5a9558 | 1230a1f36b91bd7ef2d57840dcfa013ca07e5a4a | refs/heads/master | 2022-12-16T00:44:36.798298 | 2018-08-17T13:23:02 | 2018-08-17T13:26:37 | 145,027,493 | 0 | 0 | null | 2022-12-08T02:46:37 | 2018-08-16T18:55:02 | Python | UTF-8 | Python | false | false | 545 | py | #!/usr/bin/env python
# pylint: disable=wrong-import-position
"""
Script loading tags from JSON file to data
Invocation (from project root)
./scripts/load_tags.py
"""
import os
import sys
import gflags
sys.path.append(os.getcwd())
from scripts.main_gflag import main_gflagged
from tagger.load_json import insert_tags
FLAGS = gflags.FLAGS
gflags.DEFINE_string(
"datafile", None, "file holding json data"
)
def main():
insert_tags(FLAGS.datafile)
if __name__ == "__main__":
sys.exit(main_gflagged(sys.argv, main))
| [
"gordon@practicalhorseshoeing.com"
] | gordon@practicalhorseshoeing.com |
7aaee542c16692bd7e6bafbc73a9debf04054e3d | 504dbb060f00d373278f2f210af39fb89a27916b | /karyawan.py | 5dad1d91ebbe692dd9127b5753080694a65460bd | [] | no_license | Aldidwi53/projek-PBO | e56f1b764c632f6d59a550feadac3583b35e1345 | 7232feb29e5e81e4e82241105bfd41b85ec6d21b | refs/heads/main | 2023-02-06T02:01:53.925205 | 2021-01-02T09:09:28 | 2021-01-02T09:09:28 | 324,373,345 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 206 | py | import user
import sqlite3
class Karyawan(user.User):
def __init__(self, email, password, nama, gender, alamat, telepon):
super().__init__(email, password, nama, gender, alamat, telepon) | [
"noreply@github.com"
] | Aldidwi53.noreply@github.com |
d34828fbd987211cc81fd1989af985a0e9374a74 | 128104139a52489f21df67918fae50b647f27cbe | /printswapneg.py | e3d6c3796765ea2755a55619b285b3480622b3fb | [] | no_license | AdamRichey/Python | aabb254742cf46c4d958b8aa55c26bdca100cdde | 00f1be1e235f6f791c624e3496b78d8863355e81 | refs/heads/master | 2020-03-23T05:54:10.976815 | 2018-07-16T18:15:12 | 2018-07-16T18:15:12 | 141,174,873 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 134 | py | arr=[-1]
def neg(arr):
for i in arr:
if arr[i]<0:
arr[i]="dojo"
print arr
print neg(arr) | [
"adamrichey88@gmail.com"
] | adamrichey88@gmail.com |
670f081b1ef8d14da851e983a53c42b304f46728 | ace2dc6096eb0b7a540f28e57df8459adafad6ed | /Advanced Algorithms and Complexity/Programming-Assignment-4/circuit_design/circuit_design.py | e2ec8b4f2aabddcc3f0a3899603eacd626bc5794 | [] | no_license | tdslivensky/AlgorithmsAndDataStructures | 6ad2c28204600b1f8f72228c13d29d2c3c9437c9 | e8b1011ab5210bc52854f911e2a7e41a83b36740 | refs/heads/master | 2023-01-11T16:32:49.399654 | 2020-11-13T13:49:18 | 2020-11-13T13:49:18 | 289,050,279 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 1,016 | py | # python3
n, m = map(int, input().split())
clauses = [ list(map(int, input().split())) for i in range(m) ]
# This solution tries all possible 2^n variable assignments.
# It is too slow to pass the problem.
# Implement a more efficient algorithm here.
def isSatisfiable():
for mask in range(1<<n):
result = [ (mask >> i) & 1 for i in range(n) ]
formulaIsSatisfied = True
for clause in clauses:
clauseIsSatisfied = False
if result[abs(clause[0]) - 1] == (clause[0] < 0):
clauseIsSatisfied = True
if result[abs(clause[1]) - 1] == (clause[1] < 0):
clauseIsSatisfied = True
if not clauseIsSatisfied:
formulaIsSatisfied = False
break
if formulaIsSatisfied:
return result
return None
result = isSatisfiable()
if result is None:
print("UNSATISFIABLE")
else:
print("SATISFIABLE")
print(" ".join(str(-i-1 if result[i] else i+1) for i in range(n)))
| [
"tdslivensky@gmail.com"
] | tdslivensky@gmail.com |
2a4a81a565fab19cc75a574eb4d85c9994bb0767 | c67f2d0677f8870bc1d970891bbe31345ea55ce2 | /zippy/lib-python/3/test/test_file.py | e9a1ceeeb8a48606fa1ad65140adac2fd3689d05 | [
"BSD-3-Clause"
] | permissive | securesystemslab/zippy | a5a1ecf5c688504d8d16128ce901406ffd6f32c2 | ff0e84ac99442c2c55fe1d285332cfd4e185e089 | refs/heads/master | 2022-07-05T23:45:36.330407 | 2018-07-10T22:17:32 | 2018-07-10T22:17:32 | 67,824,983 | 324 | 27 | null | null | null | null | UTF-8 | Python | false | false | 11,928 | py | import sys
import os
import unittest
from array import array
from weakref import proxy
import io
import _pyio as pyio
from test.support import TESTFN, run_unittest, gc_collect
from collections import UserList
class AutoFileTests(unittest.TestCase):
# file tests for which a test file is automatically set up
def setUp(self):
self.f = self.open(TESTFN, 'wb')
def tearDown(self):
if self.f:
self.f.close()
os.remove(TESTFN)
def testWeakRefs(self):
# verify weak references
p = proxy(self.f)
p.write(b'teststring')
self.assertEqual(self.f.tell(), p.tell())
self.f.close()
self.f = None
gc_collect()
self.assertRaises(ReferenceError, getattr, p, 'tell')
def testAttributes(self):
# verify expected attributes exist
f = self.f
f.name # merely shouldn't blow up
f.mode # ditto
f.closed # ditto
def testReadinto(self):
# verify readinto
self.f.write(b'12')
self.f.close()
a = array('b', b'x'*10)
self.f = self.open(TESTFN, 'rb')
n = self.f.readinto(a)
self.assertEqual(b'12', a.tobytes()[:n])
def testReadinto_text(self):
# verify readinto refuses text files
a = array('b', b'x'*10)
self.f.close()
self.f = self.open(TESTFN, 'r')
if hasattr(self.f, "readinto"):
self.assertRaises(TypeError, self.f.readinto, a)
def testWritelinesUserList(self):
# verify writelines with instance sequence
l = UserList([b'1', b'2'])
self.f.writelines(l)
self.f.close()
self.f = self.open(TESTFN, 'rb')
buf = self.f.read()
self.assertEqual(buf, b'12')
def testWritelinesIntegers(self):
# verify writelines with integers
self.assertRaises(TypeError, self.f.writelines, [1, 2, 3])
def testWritelinesIntegersUserList(self):
# verify writelines with integers in UserList
l = UserList([1,2,3])
self.assertRaises(TypeError, self.f.writelines, l)
def testWritelinesNonString(self):
# verify writelines with non-string object
class NonString:
pass
self.assertRaises(TypeError, self.f.writelines,
[NonString(), NonString()])
def testErrors(self):
f = self.f
self.assertEqual(f.name, TESTFN)
self.assertTrue(not f.isatty())
self.assertTrue(not f.closed)
if hasattr(f, "readinto"):
self.assertRaises((IOError, TypeError), f.readinto, "")
f.close()
self.assertTrue(f.closed)
def testMethods(self):
methods = [('fileno', ()),
('flush', ()),
('isatty', ()),
('__next__', ()),
('read', ()),
('write', (b"",)),
('readline', ()),
('readlines', ()),
('seek', (0,)),
('tell', ()),
('write', (b"",)),
('writelines', ([],)),
('__iter__', ()),
]
methods.append(('truncate', ()))
# __exit__ should close the file
self.f.__exit__(None, None, None)
self.assertTrue(self.f.closed)
for methodname, args in methods:
method = getattr(self.f, methodname)
# should raise on closed file
self.assertRaises(ValueError, method, *args)
# file is closed, __exit__ shouldn't do anything
self.assertEqual(self.f.__exit__(None, None, None), None)
# it must also return None if an exception was given
try:
1/0
except:
self.assertEqual(self.f.__exit__(*sys.exc_info()), None)
def testReadWhenWriting(self):
self.assertRaises(IOError, self.f.read)
class CAutoFileTests(AutoFileTests):
open = io.open
class PyAutoFileTests(AutoFileTests):
open = staticmethod(pyio.open)
class OtherFileTests(unittest.TestCase):
def testModeStrings(self):
# check invalid mode strings
for mode in ("", "aU", "wU+"):
try:
f = self.open(TESTFN, mode)
except ValueError:
pass
else:
f.close()
self.fail('%r is an invalid file mode' % mode)
def testStdin(self):
# This causes the interpreter to exit on OSF1 v5.1.
if sys.platform != 'osf1V5':
self.assertRaises((IOError, ValueError), sys.stdin.seek, -1)
else:
print((
' Skipping sys.stdin.seek(-1), it may crash the interpreter.'
' Test manually.'), file=sys.__stdout__)
self.assertRaises((IOError, ValueError), sys.stdin.truncate)
def testBadModeArgument(self):
# verify that we get a sensible error message for bad mode argument
bad_mode = "qwerty"
try:
f = self.open(TESTFN, bad_mode)
except ValueError as msg:
if msg.args[0] != 0:
s = str(msg)
if TESTFN in s or bad_mode not in s:
self.fail("bad error message for invalid mode: %s" % s)
# if msg.args[0] == 0, we're probably on Windows where there may be
# no obvious way to discover why open() failed.
else:
f.close()
self.fail("no error for invalid mode: %s" % bad_mode)
def testSetBufferSize(self):
# make sure that explicitly setting the buffer size doesn't cause
# misbehaviour especially with repeated close() calls
for s in (-1, 0, 1, 512):
try:
f = self.open(TESTFN, 'wb', s)
f.write(str(s).encode("ascii"))
f.close()
f.close()
f = self.open(TESTFN, 'rb', s)
d = int(f.read().decode("ascii"))
f.close()
f.close()
except IOError as msg:
self.fail('error setting buffer size %d: %s' % (s, str(msg)))
self.assertEqual(d, s)
def testTruncateOnWindows(self):
# SF bug <http://www.python.org/sf/801631>
# "file.truncate fault on windows"
os.unlink(TESTFN)
f = self.open(TESTFN, 'wb')
try:
f.write(b'12345678901') # 11 bytes
f.close()
f = self.open(TESTFN,'rb+')
data = f.read(5)
if data != b'12345':
self.fail("Read on file opened for update failed %r" % data)
if f.tell() != 5:
self.fail("File pos after read wrong %d" % f.tell())
f.truncate()
if f.tell() != 5:
self.fail("File pos after ftruncate wrong %d" % f.tell())
f.close()
size = os.path.getsize(TESTFN)
if size != 5:
self.fail("File size after ftruncate wrong %d" % size)
finally:
f.close()
os.unlink(TESTFN)
def testIteration(self):
# Test the complex interaction when mixing file-iteration and the
# various read* methods.
dataoffset = 16384
filler = b"ham\n"
assert not dataoffset % len(filler), \
"dataoffset must be multiple of len(filler)"
nchunks = dataoffset // len(filler)
testlines = [
b"spam, spam and eggs\n",
b"eggs, spam, ham and spam\n",
b"saussages, spam, spam and eggs\n",
b"spam, ham, spam and eggs\n",
b"spam, spam, spam, spam, spam, ham, spam\n",
b"wonderful spaaaaaam.\n"
]
methods = [("readline", ()), ("read", ()), ("readlines", ()),
("readinto", (array("b", b" "*100),))]
try:
# Prepare the testfile
bag = self.open(TESTFN, "wb")
bag.write(filler * nchunks)
bag.writelines(testlines)
bag.close()
# Test for appropriate errors mixing read* and iteration
for methodname, args in methods:
f = self.open(TESTFN, 'rb')
if next(f) != filler:
self.fail, "Broken testfile"
meth = getattr(f, methodname)
meth(*args) # This simply shouldn't fail
f.close()
# Test to see if harmless (by accident) mixing of read* and
# iteration still works. This depends on the size of the internal
# iteration buffer (currently 8192,) but we can test it in a
# flexible manner. Each line in the bag o' ham is 4 bytes
# ("h", "a", "m", "\n"), so 4096 lines of that should get us
# exactly on the buffer boundary for any power-of-2 buffersize
# between 4 and 16384 (inclusive).
f = self.open(TESTFN, 'rb')
for i in range(nchunks):
next(f)
testline = testlines.pop(0)
try:
line = f.readline()
except ValueError:
self.fail("readline() after next() with supposedly empty "
"iteration-buffer failed anyway")
if line != testline:
self.fail("readline() after next() with empty buffer "
"failed. Got %r, expected %r" % (line, testline))
testline = testlines.pop(0)
buf = array("b", b"\x00" * len(testline))
try:
f.readinto(buf)
except ValueError:
self.fail("readinto() after next() with supposedly empty "
"iteration-buffer failed anyway")
line = buf.tobytes()
if line != testline:
self.fail("readinto() after next() with empty buffer "
"failed. Got %r, expected %r" % (line, testline))
testline = testlines.pop(0)
try:
line = f.read(len(testline))
except ValueError:
self.fail("read() after next() with supposedly empty "
"iteration-buffer failed anyway")
if line != testline:
self.fail("read() after next() with empty buffer "
"failed. Got %r, expected %r" % (line, testline))
try:
lines = f.readlines()
except ValueError:
self.fail("readlines() after next() with supposedly empty "
"iteration-buffer failed anyway")
if lines != testlines:
self.fail("readlines() after next() with empty buffer "
"failed. Got %r, expected %r" % (line, testline))
f.close()
# Reading after iteration hit EOF shouldn't hurt either
f = self.open(TESTFN, 'rb')
try:
for line in f:
pass
try:
f.readline()
f.readinto(buf)
f.read()
f.readlines()
except ValueError:
self.fail("read* failed after next() consumed file")
finally:
f.close()
finally:
os.unlink(TESTFN)
class COtherFileTests(OtherFileTests):
open = io.open
class PyOtherFileTests(OtherFileTests):
open = staticmethod(pyio.open)
def test_main():
# Historically, these tests have been sloppy about removing TESTFN.
# So get rid of it no matter what.
try:
run_unittest(CAutoFileTests, PyAutoFileTests,
COtherFileTests, PyOtherFileTests)
finally:
if os.path.exists(TESTFN):
os.unlink(TESTFN)
if __name__ == '__main__':
test_main()
| [
"thezhangwei@gmail.com"
] | thezhangwei@gmail.com |
0b8e26a5b14106ea4c7a0fcf55baed32e47f43de | 06677c398ea51e3bf78dc45db8cdf97a2b2a296b | /table-cards/cardgen.py | 773492e4f3acd63bb1b3f1079022250713136a9b | [] | no_license | DoESLiverpool/Liverpool-Makefest-2017 | ac0e9df9a8b7fdb68d70d81985690ce25192d856 | 5590cb5f58a3aa403d49c54a6e25ec5e85358b10 | refs/heads/master | 2021-01-21T19:51:51.046353 | 2019-06-28T14:48:22 | 2019-06-28T14:48:22 | 92,169,342 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,995 | py | #!/usr/bin/env python3
import csv, re, os, shutil, argparse, urllib
#import lxml.etree.ElementTree as ET
from lxml import etree
SVG_FOLDER = 'svgs'
PDF_FOLDER = 'pdfs'
def main():
# parse some flags here
parser = argparse.ArgumentParser()
parser.add_argument("-i", "--infile", help="ods or xlsx input file", default='makers.ods')
parser.add_argument("-t", "--template", help="SVG template filename", default='template.svg')
parser.add_argument("-o", "--outfile", help="PDF outfile name", default='makers.pdf')
parser.add_argument("-c", "--keepcsv", help="Keep the temporary CSV file", default=False, action='store_true')
parser.add_argument("-s", "--keepsvgs", help="Keep temporary SVG files", default=False, action='store_true')
parser.add_argument("-p", "--keeppdfs", help="Keep temporary PDF files", default=False, action='store_true')
parser.add_argument("-x", "--debugflag", help="process one entry and stop", default=False, action='store_true')
args = parser.parse_args()
infile = args.infile
template = args.template
outfile = args.outfile
keep_pdfs = args.keeppdfs
keep_svgs = args.keepsvgs
keep_csv = args.keepcsv
debugflag = args.debugflag
# generate the csv file (comma separated, double quotetd, utf-8)
# TODO: check if libreoffice is running, otherwise this generation fails silently
# because the lockfile exists
os.system('libreoffice --headless --convert-to csv:"Text - txt - csv (StarCalc)":44,34,76,1,1 --outdir . ' + infile)
csvfilename = re.sub(r'\.[a-zA-Z0-9]+$', '', infile) + '.csv'
# check the required dirs exist
if not os.path.exists(SVG_FOLDER):
os.makedirs(SVG_FOLDER)
if not os.path.exists(PDF_FOLDER):
os.makedirs(PDF_FOLDER)
# create each file from line of csv file
with open(csvfilename) as csvfile:
reader = csv.DictReader(csvfile)
i = 1
for row in reader:
#print(row.keys())
# generate the required variables to substitute into the SVG
filesafe_name = re.sub(r"[^\w\s]", '', row['{title}'])
filesafe_name = re.sub(r"\s+", '-', filesafe_name)
filesafe_name = str(i).zfill(2) + '-' + filesafe_name.strip()
filesafe_name = (filesafe_name[:14]) if len(filesafe_name) > 14 else filesafe_name
title = row['{title}'].strip()
name = row['{name}'].strip()
description = row['{description}'].replace('_x000D_','')
# standardise *some* of the possible twitter and web inputs
twitter = '@' + row['{twitter}'].strip().replace('http://','').replace('https://','').replace('twitter.com/','').lstrip('@').strip()
website = row['{website}'].strip().replace('http://','').replace('https://','').replace('www.','').strip()
# parse vars to standardise text input
# replace the placeholders in the new file
svg_file = SVG_FOLDER + '/' + filesafe_name + '.svg'
# read the svg template file in
#tree = ET.parse(template)
#root = tree.getroot()
tree = etree.parse(template)
root = tree.getroot()
for para in root.findall('.//{http://www.w3.org/2000/svg}flowPara'):
if para.text == '{title}':
para.text = title
if len(title) >= 34:
# reduce the text size
parent = para.find('..')
style_tag = parent.attrib['style']
# find the current font size
font_size_tag = re.search('font-size:[0-9.]+px;', style_tag).group()
font_size = float(re.search(r'[0-9.]+', font_size_tag).group())
if len(title) >= 50:
font_size = font_size*0.75
else:
font_size = font_size*0.85
style_tag = re.sub(r'font-size:[0-9.]+px;', 'font-size:' + str(font_size) + 'px;', style_tag)
parent.attrib['style'] = style_tag
print('title font-size: ' + str(font_size) + ' px;')
#print(parent.attrib['style'])
elif para.text == '{name}':
para.text = name
elif para.text == '{description}':
para.text = description
if len(description) >= 512:
# reduce the text size
parent = para.find('..')
style_tag = parent.attrib['style']
# find the current font size
font_size_tag = re.search('font-size:[0-9.]+px;', style_tag).group()
font_size = float(re.search(r'[0-9.]+', font_size_tag).group())
if len(description) > 1200:
font_size = font_size*0.65
elif len(description) > 800:
font_size = font_size*0.75
else:
font_size = font_size*0.85
style_tag = re.sub(r'font-size:[0-9.]+px', 'font-size:' + str(font_size) + 'px', style_tag)
parent.attrib['style'] = style_tag
print('description font-size: ' + str(font_size) + ' px;')
elif para.text == '{twitter}':
if twitter != '@': # is empty
para.text = twitter
else:
para.text = ''
elif para.text == '{website}':
if website[-1:] == '/':
para.text = website[:-1]
else:
para.text = website
# write the adjusted svg
tree.write(svg_file)
pdf_file = PDF_FOLDER + '/' + filesafe_name + '.pdf'
os.system('inkscape --without-gui --file ' + svg_file + ' --export-text-to-path --export-area-page --export-pdf ' + pdf_file)
print('Created: ' + title)
i+=1
if debugflag == True:
print ('Filename: ' + filesafe_name)
print ('Name: ' + name)
print ('Title: ' + title + ' [' + str(len(title)) + ']')
print ('Description: ' + description + ' [' + str(len(description)) + ']')
print ('Twitter: ' + twitter)
print ('Website: ' + website)
quit()
# concatenate all the pdf pages
os.chdir(PDF_FOLDER)
os.system('pdftk *.pdf output ../' + outfile)
os.chdir('..')
# cleanup temporary files
if keep_csv == False:
os.remove(csvfilename)
if keep_svgs == False:
shutil.rmtree(SVG_FOLDER)
if keep_pdfs == False:
shutil.rmtree(PDF_FOLDER)
if __name__ == "__main__":
main()
| [
"patrick@defproc.co.uk"
] | patrick@defproc.co.uk |
81adc9b89c325fae8eb969a4530b965c9f2ee337 | 83de24182a7af33c43ee340b57755e73275149ae | /aliyun-python-sdk-ice/aliyunsdkice/request/v20201109/DescribeQueryConfigsRequest.py | d4de87a4f55304fb57408fb816c6f72a3b5b2c81 | [
"Apache-2.0"
] | permissive | aliyun/aliyun-openapi-python-sdk | 4436ca6c57190ceadbc80f0b1c35b1ab13c00c7f | 83fd547946fd6772cf26f338d9653f4316c81d3c | refs/heads/master | 2023-08-04T12:32:57.028821 | 2023-08-04T06:00:29 | 2023-08-04T06:00:29 | 39,558,861 | 1,080 | 721 | NOASSERTION | 2023-09-14T08:51:06 | 2015-07-23T09:39:45 | Python | UTF-8 | Python | false | false | 1,437 | py | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from aliyunsdkcore.request import RpcRequest
from aliyunsdkice.endpoint import endpoint_data
class DescribeQueryConfigsRequest(RpcRequest):
def __init__(self):
RpcRequest.__init__(self, 'ICE', '2020-11-09', 'DescribeQueryConfigs','ice')
self.set_method('POST')
if hasattr(self, "endpoint_map"):
setattr(self, "endpoint_map", endpoint_data.getEndpointMap())
if hasattr(self, "endpoint_regional"):
setattr(self, "endpoint_regional", endpoint_data.getEndpointRegional())
def get_Type(self): # String
return self.get_query_params().get('Type')
def set_Type(self, Type): # String
self.add_query_param('Type', Type)
| [
"sdk-team@alibabacloud.com"
] | sdk-team@alibabacloud.com |
d95df558fb04770dfcce9d62774614c18e3bfa25 | 401e455157f74b28b0ad12df6bdae3c159888db6 | /tiviapp/apis/twitter_rest.py | 2711d64b25094018a1871401c9eeafc7e1407e66 | [] | no_license | AnilSener/tivi | 0624e356524ae03fddd771a43a174211b8e1909d | 701e1ab2948d898fe12999bc5be954f61a3f4e99 | refs/heads/master | 2021-01-01T06:04:31.707925 | 2015-07-03T15:34:36 | 2015-07-03T15:34:36 | 38,221,946 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,230 | py | __author__ = 'anil'
from twython import TwythonStreamer,Twython,TwythonError
from tivi.celery import app
from tiviapp.models import *
import time
####################################################################
consumer_key="Vs7V2k4vPWMMyTFqLzqPkM6wE"
consumer_secret="aWNRzh74LUT1fuW35y6VzRDtvuimQ4LjFGMnMMkEXI0Y9LSpkf"
access_token="258113369-63Y2Cqr9q0Bo02WU4AS8Bjiv3JnHP2Us7HimK26G"
access_token_secret="Z4Sf9EyLbOJ4jPI5WlZPZUyv3OwluuZXiKXn0pamk8Dly"
###################################################################
twitter = Twython(consumer_key, consumer_secret,access_token,access_token_secret)
@app.task()
def exec_User_Follows():
twitter_users=TwitterUser.objects.all()
if len(twitter_users)==0:
print "No users available Wait 5 minutes for the next API call"
time.sleep(300)
else:
for i,user in enumerate(twitter_users):
print user.userName,"!!!"
try:
print "!!!TIME FOR FOLLOWERS!!!"
followers=twitter.get_followers_list(screen_name=user.userName,include_user_entities=True,count=200)
for f in followers:
print f
except TwythonError as e:
print e.message | [
"anil_sener@yahoo.com"
] | anil_sener@yahoo.com |
d01f9d1b57765a72c85ec040eab037e9d12c89bb | ca77e9e45d666771c7b0897e7e3093b3d3c12f65 | /scripts/trigger/add_prices.py | ec79a8be575fe0f59c9b16754b18afc1910a7a29 | [] | no_license | 2gDigitalPost/custom | 46175d3a3fc4c3be21dc20203ff0a48fb93b5639 | 6a3a804ef4ef6178044b70ad1e4bc5c56ab42d8d | refs/heads/master | 2020-04-04T07:40:17.962611 | 2016-12-28T18:35:28 | 2016-12-28T18:35:28 | 39,648,283 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,857 | py | """
This file was generated automatically from a custom script found in Project -> Script Editor.
The custom script was moved to a file so that it could be integrated with GitHub.
"""
__author__ = 'Topher.Hughes'
__date__ = '04/08/2015'
import traceback
def main(server=None, input=None):
"""
The main function of the custom script. The entire script was copied
and pasted into the body of the try statement in order to add some
error handling. It's all legacy code, so edit with caution.
:param server: the TacticServerStub object
:param input: a dict with data like like search_key, search_type, sobject, and update_data
:return: None
"""
if not input:
input = {}
try:
# CUSTOM_SCRIPT00035
# Matthew Tyler Misenhimer
# This is used to have the prices on projects trickle up to titles, then orders
# This is DEPRECATED
sobj = input.get('sobject')
sk = input.get('search_key')
price_str = sobj.get('price')
price = 0
if price_str not in [None,'']:
price = float(price_str)
proj = server.eval("@SOBJECT(twog/proj['code','%s'])" % sobj.get('proj_code'))[0]
current_proj_price_str = proj.get('price')
current_proj_price = 0
if current_proj_price_str not in [None,'']:
current_proj_price = float(current_proj_price_str)
new_proj_price = current_proj_price + price
server.update(proj.get('__search_key__'), {'price': new_proj_price})
title = server.eval("@SOBJECT(twog/title['code','%s'])" % proj.get('title_code'))[0]
current_title_price_str = title.get('price')
current_title_price = 0
if current_title_price_str not in [None,'']:
current_title_price = float(current_title_price_str)
new_title_price = current_title_price + price
server.update(title.get('__search_key__'), {'price': new_title_price})
order = server.eval("@SOBJECT(twog/order['code','%s'])" % title.get('order_code'))[0]
current_order_price_str = order.get('price')
current_order_price = 0
if current_order_price_str not in [None,'']:
current_order_price = float(current_order_price_str)
new_order_price = current_order_price + price
server.update(order.get('__search_key__'), {'price': new_order_price})
except AttributeError as e:
traceback.print_exc()
print str(e) + '\nMost likely the server object does not exist.'
raise e
except KeyError as e:
traceback.print_exc()
print str(e) + '\nMost likely the input dictionary does not exist.'
raise e
except Exception as e:
traceback.print_exc()
print str(e)
raise e
if __name__ == '__main__':
main()
| [
"topher.hughes@2gdigital.com"
] | topher.hughes@2gdigital.com |
3010b06f90027e9d8277db75879d2a23679e061c | 934107eaba17b352bf7bf3a9c0a45af4f263fd54 | /favorites/manager.py | 5106bb9da8161cbc63e045832e3288fd4baca22e | [] | no_license | quentin338/Purebeurre-p8 | d0324f0fbc7e96a1418b367ea41ef3f4f51cc437 | 15bb4192df331d790ef28140e65213ac604cf96e | refs/heads/master | 2023-08-07T22:25:10.975870 | 2023-07-26T10:21:04 | 2023-07-26T10:21:04 | 207,123,046 | 0 | 0 | null | 2023-07-26T10:21:06 | 2019-09-08T14:17:16 | CSS | UTF-8 | Python | false | false | 310 | py | from django.db import models
from django.db.utils import IntegrityError
class FavoriteManager(models.Manager):
def is_favorite(self, user, ancient_product, new_product):
return bool(self.filter(user=user, ancient_product=ancient_product,
new_product=new_product))
| [
"quentin.bertrand@yahoo.fr"
] | quentin.bertrand@yahoo.fr |
14450e65ad686acdab9fb6fecaa1b50a8a7d5106 | 84fda250fd32b37d74f07f6d00106226881e70ee | /shop/cart/admin.py | 85a9d7c503ea1cc245e73f3574a659d5c5459871 | [] | no_license | NovosadVictor/OnlineShop | 8b80b1122f9d8019fb2d1641e08ea9eb64cc6ed6 | afe6826bf0494be6af3a033cf8c2e3a205ffb571 | refs/heads/master | 2020-12-02T16:13:26.312803 | 2017-08-22T16:29:55 | 2017-08-22T16:29:55 | 96,091,899 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 255 | py | from django.contrib import admin
from .models import ProductInCart
class ProductInCartAdmin(admin.ModelAdmin):
list_display = ['owner', 'product', 'quantity',]
list_editable = ['quantity',]
admin.site.register(ProductInCart, ProductInCartAdmin)
| [
"novosad_msu@mail.ru"
] | novosad_msu@mail.ru |
1a7e98eac3530ccd63f562cdfb6de3ad851647f7 | 5223229cbdbe883c6c1c09980c60d845e8255dd3 | /laliga_sb_analysis/Scripts/Barca_Manager_tenure_graph.py | e6efd1c05d1abfb8fc9f6185443ca62b15abc13d | [] | no_license | derrik-hanson/Python_Analysis_xG_Barca_plus | 39b112c47672b5f3133a1b439c942e4cee35791e | 65af5acb02deadb610b182a4b74b444114eb3ec6 | refs/heads/main | 2023-08-26T18:18:17.546730 | 2021-11-13T06:26:36 | 2021-11-13T06:26:36 | 422,980,924 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,627 | py | #!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Fri Nov 12 19:39:17 2021
@author: Derrik Hanson
"""
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
"""
Created on Fri Nov 12 17:03:57 2021
@author: Derrik Hanson
"""
import plotly.express as px
import pandas as pd
manager_tenures = [
{'manager_name': 'Frank Rijkaard', 'start':'2003-06' ,'end':'2008-06'},
{'manager_name': 'Pep Guardiola', 'start':'2008-06' ,'end':'2012-06'},
{'manager_name': 'Tito Vilanova', 'start':'2012-07' ,'end':'2013-01'},
{'manager_name': 'Jordi Roura', 'start':'2013-01' ,'end':'2013-03'},
{'manager_name': 'Tito Vilanova', 'start':'2013-03' ,'end':'2013-07'},
{'manager_name': 'Gerard Martino', 'start':'2013-07' ,'end':'2014-05'},
{'manager_name': 'Luis Enrique', 'start':'2014-05' ,'end':'2017-05'},
{'manager_name': 'Ernesto Valverde', 'start':'2017-05' ,'end':'2020-01'},
{'manager_name': 'Quique Setien', 'start':'2020-01' ,'end':'2020-08'},
{'manager_name': 'Ronald Koeman', 'start':'2020-08' ,'end':'2021-10'},
]
# load DataFrame
df = pd.DataFrame(manager_tenures)
# Create Gantt Plot
fig = px.timeline(df, x_start="start", x_end="end", y="manager_name",
labels = {
'manager_name': 'Manager Name'}
)
fig.update_layout(
title={
'text': "Barcelona Manager Tenures",
'y':0.95,
'x':0.5,
'xanchor': 'center',
'yanchor': 'top'})
fig.update_yaxes(autorange="reversed") # otherwise tasks are slisted from the bottom up
fig.show()
fig.write_image("figures/barca_manager_tenure.pdf") | [
"hanson.derrik@gmail.com"
] | hanson.derrik@gmail.com |
de9ceaa3537c1f1edf2a30fedb2a4f538e0eec02 | a2efa9d89a721aae5016280ca166caffab97e94f | /exercices_EDX_W1.py | 44e668445b4b6b51f8543dd707285978c0d16491 | [] | no_license | samthib/python_edx | 3dc8378c92092eedfb70b0890ee801465e3a4cd9 | 2970745f6671a0d314b2cd32261928c7a55da3f9 | refs/heads/master | 2022-07-04T19:00:19.037593 | 2020-05-05T20:54:57 | 2020-05-05T20:54:57 | 258,567,195 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,555 | py | # -*- coding: utf-8 -*-
"""
Created on Mon Apr 20 21:10:42 2020
@author: Sam
"""
##--------------------------------##
#Exercice 1
import string
alphabet = string.ascii_letters
sentence = 'Jim quickly realized that the beautiful gowns are expensive'
count_letters = {}
for i in range(len(sentence)):
if sentence[i] in count_letters:
count_letters[sentence[i]] += 1
else:
count_letters[sentence[i]] = 1
def counter(input_string):
alphabet = string.ascii_letters
for i in range(len(input_string)):
if input_string[i] in alphabet:
if input_string[i] in count_letters:
count_letters[input_string[i]] += 1
else:
count_letters[input_string[i]] = 1
return count_letters
#print(counter(sentence))
address = """Four score and seven years ago our fathers brought forth on this continent, a new nation,
conceived in Liberty, and dedicated to the proposition that all men are created equal. Now we are engaged in a
great civil war, testing whether that nation, or any nation so conceived and so dedicated, can long endure.
We are met on a great battle-field of that war. We have come to dedicate a portion of that field, as a final
resting place for those who here gave their lives that that nation might live. It is altogether fitting and proper
that we should do this. But, in a larger sense, we can not dedicate -- we can not consecrate -- we can not hallow --
this ground. The brave men, living and dead, who struggled here, have consecrated it, far above our poor power to add
or detract. The world will little note, nor long remember what we say here, but it can never forget what they did here.
It is for us the living, rather, to be dedicated here to the unfinished work which they who fought here have thus far so
nobly advanced. It is rather for us to be here dedicated to the great task remaining before us -- that from these honored
dead we take increased devotion to that cause for which they gave the last full measure of devotion -- that we here
highly resolve that these dead shall not have died in vain -- that this nation, under God, shall have a new birth of
freedom -- and that government of the people, by the people, for the people, shall not perish from the earth."""
address_count = counter(address)
#print(address_count)
count_letters_max = max(address_count, key=address_count.get)
#print(count_letters_max)
##--------------------------------##
#Excercie 2
import math
#print(math.pi/4)
import random
random.seed(1) # Fixes the see of the random number generator.
def rand():
return random.uniform(-1, 1)
rand()
def distance(x, y):
diff_x = y[0]-x[0]
diff_y = y[1]-x[1]
distance = math.sqrt(diff_x**2 + diff_y**2)
return distance
x=(0,0)
y=(1,1)
distance(x, y)
def in_circle(x, origin = [0,0]):
radius = distance(x, origin)
if radius < 1:
return True
else:
return False
in_circle((1,1))
random.seed(1)
R=10000
inside = []
count_true = 0
for i in range(R):
point = in_circle((rand(),rand()))
inside.append(point)
if point:
count_true += 1
#print(count_true / R)
difference = (math.pi / 4) - (count_true / R)
#print(difference)
##--------------------------------##
# Exercice 3
"""
Corection
def moving_window_average(x, n_neighbors=1):
n = len(x)
width = n_neighbors*2 + 1
x = [x[0]]*n_neighbors + x + [x[-1]]*n_neighbors
return [sum(x[i:(i+width)]) / width for i in range(n)]
x = [0,10,5,3,1,5]
#print(sum(moving_window_average(x, 1)))
"""
def moving_window_average(x, n_neighbors=1):
width = n_neighbors*2 + 1
x = [x[0]]*n_neighbors + x + [x[-1]]*n_neighbors
n = len(x)
list_x=[]
for i in range(n_neighbors,n-n_neighbors):
sum_x=0
for j in range(-n_neighbors, n_neighbors+1):
sum_x = sum_x + x[i+j]
mean = sum_x / width
list_x.append(mean)
return list_x
x = [0,10,5,3,1,5]
#print(moving_window_average(x, 1))
#print(sum(moving_window_average(x, 1)))
R = 1000
Y = []
x = []
ranges = []
random.seed(1)
for i in range(R):
x.append(random.uniform(0,1))
for i in range(1, 10):
Y.append(moving_window_average(x, i))
for i in range(9):
ranges.append(max(Y[i])-min(Y[i]))
#print(Y[5][9])
#print(ranges)
| [
"noreply@github.com"
] | samthib.noreply@github.com |
2566347edc4da7664cffa99c4c72f58678e8c26a | 648384ac8ff1e1a414c41c1ed19c6d54d7f1aeb2 | /pygame_basic/4_keyboard_event.py | 98eb5facda5d2b3900695aa414f57d47e2adf207 | [] | no_license | Online-abayss/-- | dadb65b372ed7c39c391bc0a85355a314028ce19 | 4ac61ad20167a784419365317d61722d5bbddf3b | refs/heads/main | 2023-08-03T13:13:51.623181 | 2021-09-23T08:42:10 | 2021-09-23T08:42:10 | 397,529,231 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,212 | py | import pygame
from pygame.constants import K_LEFT, K_RIGHT
pygame.init() # 초가화작업 (무조건 필수) (클래스 정의할떄도 self.init 하는것처럼 그런듯)
# 화면 크기 설정
screen_width = 480 ## 가로크기
screen_height = 640 ## 세로크기
screen = pygame.display.set_mode((screen_width,screen_height)) ## 게임 화면 크기 설정
# 화면 타이틀 설정 (2. background 시작)
pygame.display.set_caption("Test Game") # 게임 타이틀 제작
# 배경 이미지 불러오기
background = pygame.image.load("C:\\Users\\kang\\Desktop\\Pythonworkspace\\pygame_basic\\background.png")
# 캐릭터(스프라이트) 불러오기 (3.man sprite 시작)
character = pygame.image.load("C:\\Users\\kang\\Desktop\\Pythonworkspace\\pygame_basic\\character.png")
character_size = character.get_rect().size ## 캐릭터의 이미지의 가로 및 세로 크기값을 알수있음.
character_width = character_size[0] # 캐릭터의 가로 크기
character_heigth = character_size[1] # 캐릭터의 세로 크기
# 캐릭터 움직임의 관한 좌표를 설정
# 좌표는 11시 꼭짓점 기준으로 0,0을 잡고 우측 밑으로 증가한다.
# y좌표를 바로 밑에껏처럼 하면 캐릭터가 안보인다. 왜냐하면 캐릭터도 마찬가지로 좌표는 11시 꼭짓점을 기준으로 잡아주기에
# 캐릭터의 크기를 생각하고 그만큼 위로 올려서 보이게 해야한다. 또한 중앙으로 캐릭터를 옮기고 싶으면 그냥 화면 가로/2가 아닌 캐릭터 크기의 절반만큼 더 왼쪽으로 옮겨야한다.
character_x_pos = (screen_width/2) -(character_width/2)# x위치를 설정
character_y_pos = screen_height - character_heigth# Y위치를 설정
# 이동 할 좌표
to_x = 0
to_y = 0
# 이벤트 루프
# 키보드 입력에 따른 이동 여부 설정(4. keyboard_event 시작)
running = True # 게임이 계속 진행중인지? 파악
while running:
for event in pygame.event.get(): # 키보드 및 마우스 입력이 들어올경우 그 값에 대응으로 처리 (이벤트 발생 여부)
if event.type == pygame.QUIT: ## 1시 방향 X 표시의 창끄기 표시 명령어
running = False ## 내가 실수로 = 한개만 할걸 두개로 해서 확정이 아닌 조건으로 되서 무한루프로 빠져나오지 못했음.
if event.type == pygame.KEYDOWN: #키가 눌려졌는지 확인
if event.key == pygame.K_LEFT: # 캐릭터를 좌측으로 이동
to_x -= 2
elif event.key == pygame.K_RIGHT: # 캐릭터를 우측으로 이동
to_x += 2
elif event.key == pygame.K_UP: # 캐릭터를 위로 이동
to_y -= 2
elif event.key == pygame.K_DOWN: # 캐릭터를 밑으로 이동
to_y += 2
if event.type == pygame.KEYUP: # 키보드를 때면 멈추기.
if event.key == pygame.K_LEFT or event.key == pygame.K_RIGHT:
to_x = 0
elif event.key == pygame.K_UP or event.key == pygame.K_DOWN:
to_y = 0
character_x_pos += to_x
character_y_pos += to_y
# 화면 밖으로 넘어가는걸 방지
if character_x_pos < 0:
character_x_pos = 0
elif character_x_pos > screen_width - character_width: # 우측 끝 - 캐릭터 넓이만큼
character_x_pos = screen_width - character_width
if character_y_pos < 0:
character_y_pos = 0
elif character_y_pos > screen_height - character_heigth: # 스크린 맨밑 - 캐릭터 높이만큼
character_y_pos = screen_height - character_heigth
screen.blit(background, (0,0)) #배경 그리기 ## 여기까지만 하면 반영을 하지 않는다.
#rgb값을 이용하여 배경을 넣을수 있다.
#screen.fill((0,0,255)
screen.blit(character, (character_x_pos,character_y_pos)) # 캐릭터 그리기 및 위치 설정한 값으로 지정
pygame.display.update() # 매 프레임마다 배경을 그려줘야 하기에 설정함
# 게임 종료
pygame.quit()
| [
"noreply@github.com"
] | Online-abayss.noreply@github.com |
0192859b8b887bddad03294ac6c0618bf2e5e2a9 | 0ce9fbb56b1bb142eed50491c00adeb5eed52d3f | /chatbot.py | 01cda0956f43c0b5dbe7a8189c09280cd267198e | [] | no_license | lucasB97/ARTUR | 0336a7c1bc0e6fc9504dbd46a1a57e2d7c4dd299 | 287625fa5f30923abb676b6ce3fb326ff8ceda81 | refs/heads/main | 2023-04-27T12:01:47.335612 | 2021-04-17T17:31:16 | 2021-04-17T17:31:16 | 316,075,171 | 0 | 0 | null | 2021-04-17T02:59:21 | 2020-11-25T23:21:38 | HTML | UTF-8 | Python | false | false | 1,095 | py | from chatterbot import ChatBot
from chatterbot.trainers import ListTrainer
from chatterbot.trainers import ChatterBotCorpusTrainer
# Creating ChatBot Instance
chatbot = ChatBot(
'A.R.T.U.R.',
storage_adapter='chatterbot.storage.SQLStorageAdapter',
logic_adapters=[
{
'import_path': 'chatterbot.logic.BestMatch',
'default_response': 'Me desculpe, mas eu não entendi. Ainda estou aprendendo :(',
'maximum_similarity_threshold': 0.90
}
],
filters=[
'chatterbot.filters.RepetitiveResponseFilter'
],
database_uri='sqlite:///database.sqlite3'
)
# Training with Personal Ques & Ans
training_data = open('training/ques_ans.txt').read().splitlines()
trainer = ListTrainer(chatbot)
trainer.train(training_data)
# Training with Portugues Corpus Data
trainer_corpus = ChatterBotCorpusTrainer(chatbot)
trainer_corpus.train(
"chatterbot.corpus.portuguese",
"chatterbot.corpus.portuguese.greetings",
"chatterbot.corpus.portuguese.conversations",
"chatterbot.corpus.portuguese.linguistic_knowledge"
)
| [
"lucasbessa708@gmail.com"
] | lucasbessa708@gmail.com |
a9c2ba7de5529d4ee631d9a51afa34de0f801869 | 78f3ffc90eec06e3ea638b0a87b73562dc311984 | /damsht.py | 9665480edef3cb1d8c6ffd418a561d9b1c8e1296 | [] | no_license | ashavt/dsh | a19647db44ce8112ce80184c6b5351cd16b649ec | deef4281ddd71d472b4b1d0b0b633634da3bf6a5 | refs/heads/master | 2022-12-22T07:05:45.672516 | 2018-02-18T09:13:54 | 2018-02-18T09:13:54 | 121,934,464 | 0 | 0 | null | 2022-12-08T00:54:53 | 2018-02-18T08:49:08 | Python | UTF-8 | Python | false | false | 2,224 | py | import requests
import datetime
class BotHandler:
def __init__(self, token):
self.token = token
self.api_url = "https://api.telegram.org/bot{}/".format(token)
def get_updates(self, offset=None, timeout=30):
method = 'getUpdates'
params = {'timeout': timeout, 'offset': offset}
resp = requests.get(self.api_url + method, params)
result_json = resp.json()['result']
return result_json
def send_message(self, chat_id, text):
params = {'chat_id': chat_id, 'text': text}
method = 'sendMessage'
resp = requests.post(self.api_url + method, params)
return resp
def get_last_update(self):
get_result = self.get_updates()
if len(get_result) > 0:
last_update = get_result[-1]
else:
last_update = get_result[len(get_result)]
return last_update
greet_bot = BotHandler(token)
greetings = ('hello', 'hi', 'greetings', 'sup')
now = datetime.datetime.now()
def main():
new_offset = None
today = now.day
hour = now.hour
while True:
greet_bot.get_updates(new_offset)
last_update = greet_bot.get_last_update()
last_update_id = last_update['update_id']
last_chat_text = last_update['message']['text']
last_chat_id = last_update['message']['chat']['id']
last_chat_name = last_update['message']['chat']['first_name']
if last_chat_text.lower() in greetings and today == now.day and 6 <= hour < 12:
greet_bot.send_message(last_chat_id, 'Доброе утро {}'.format(last_chat_name))
today += 1
elif last_chat_text.lower() in greetings and today == now.day and 12 <= hour < 17:
greet_bot.send_message(last_chat_id, 'Добрый день {}'.format(last_chat_name))
today += 1
elif last_chat_text.lower() in greetings and today == now.day and 17 <= hour < 23:
greet_bot.send_message(last_chat_id, 'Добрый вечер {}'.format(last_chat_name))
today += 1
new_offset = last_update_id + 1
if __name__ == '__main__':
try:
main()
except KeyboardInterrupt:
exit()
| [
"ashavt@yandex.ru"
] | ashavt@yandex.ru |
6128440a21b6bad0591564200e67f0c7ab7f0018 | a219c9b0f3ccd1b35c3bb7bb3c7b50e1d9d8ef93 | /d002_1_for_dongusu.py | 9910b109c6ddd3f3d306143754c0ebd1ef137785 | [] | no_license | f0xmulder/python_ornekleri | 3293541b5d4e594dc39e6df623e47ecd4e5e94c2 | d1ebbcefdd7390a4e20a61864b150097f9919e29 | refs/heads/master | 2022-11-04T07:12:20.766931 | 2017-06-22T13:30:45 | 2017-06-22T13:30:45 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 495 | py | #!/usr/bin/env python2
# -*- coding: utf-8 -*-
"""
For Döngüsü
cok satirli
aciklama
Uşak Universitesi
"""
# tek satır aciklama
import numpy
print "Program Baslangıcı"
print "0-10 arası 1.7 artarak"
elemanlar=numpy.arange(0,10,1.7)
print "Sayı\t(3)\t(5)\t(7)"
for e in elemanlar:
print e,"\t", e%3,"\t", e%5,"\t", e%7
print "Herhangi bir liste"
elemanlar=[3,8,7.2,85]
print "Sayı\t(3)\t(5)\t(7)"
for e in elemanlar:
print e,"\t", e%3,"\t", e%5,"\t", e%7
print "Program Sonu"
| [
"noreply@github.com"
] | f0xmulder.noreply@github.com |
102f8aafd37eaff350427f12ac5e476d7dda9c04 | 992816021cb79580e9c739ca2650fe9bc83829bd | /medapp/profile/admin.py | b838f95fde2b89db4d347aec60b042db9c24c2e0 | [] | no_license | etiennekruger/medapp-api | 096de56dc523425981b48c6d80930bb3df799639 | af9232f548db28e6716c0a14e38b44dcb9b57690 | refs/heads/master | 2020-04-18T22:58:35.859694 | 2012-07-16T06:17:17 | 2012-07-16T06:17:17 | 5,150,748 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 211 | py | from django.contrib import admin
from profile.models import Profile
class ProfileAdmin(admin.ModelAdmin):
list_display = ['__unicode__', 'created', 'updated']
admin.site.register(Profile, ProfileAdmin)
| [
"makhonin@steelkiwi.com"
] | makhonin@steelkiwi.com |
7875cbdde6257275a92695eeecc06813145d79d1 | 1ad64831a79a6b163f6a31a9255994c1e4d8a322 | /settings.py | a2fe5dbf18ed9edbb73469731b1b139a097fe630 | [
"Apache-2.0"
] | permissive | benhosmer/brzr-desktop | 1f17b767a9d1b7d6f4bbb3787e5deb9df92f9165 | 1e5e88e0b33325d35002636192019a40d080224e | refs/heads/master | 2020-03-30T07:49:24.701545 | 2013-07-11T16:29:12 | 2013-07-11T16:29:12 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 184 | py | import ConfigParser
config = ConfigParser.RawConfigParser()
config.read('brzr.cfg')
database_name = config.get('main', 'database_name')
event_name = config.get('main', 'event_name')
| [
"ben.hosmer@gmail.com"
] | ben.hosmer@gmail.com |
feee22f88d28bb23ee6a97c99f3208c769d58972 | dc569d08a025447e5386abf0388c99d45bdbfcf1 | /utils.py | 6cc2d12ab1651734c79094fcbdcdd4f3dd2f23ad | [] | no_license | samuelBB/Translating-OOV-Words-Via-Images | cf992d1f27f10352fee61bc89a60ac0bc2d91a9f | 5ae8f3b89786a0813e93b619189e181de8a955dd | refs/heads/master | 2020-05-30T10:08:45.056985 | 2019-05-31T22:37:25 | 2019-05-31T22:37:25 | 189,665,688 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,088 | py | """
misc utility functions
"""
import re
import os
import errno
import pickle
import random
import logging
import argparse
import datetime as dt
RANDOM_SEED = 2018
### logging
class UpToLevel(object):
def __init__(self, lvl=logging.FATAL):
self.lvl = lvl
def filter(self, record):
return record.levelno <= self.lvl
ROOT = '*'
def init_logging(file=None, stdout=False, stderr=False,
lo_lvl=logging.DEBUG, hi_lvl=logging.FATAL,
file_lo_lvl=None, stdout_lo_lvl=None, stderr_lo_lvl=None,
file_hi_lvl=None, stdout_hi_lvl=None, stderr_hi_lvl=None,
fmt='[%(asctime)s|%(levelname)s|%(module)s'
'.%(funcName)s:%(lineno)d] %(message)s',
datefmt='%Y-%m-%d_%H:%M:%S',
mode='w'):
logger = logging.getLogger(ROOT)
if is_(lo_lvl):
logger.setLevel(lo_lvl)
if is_(hi_lvl):
logger.addFilter(UpToLevel(hi_lvl))
for name, obj, args, prefix in [
('stdout', stdout, [logging.sys.stdout], 'Stream'),
('stderr', stderr, (), 'Stream'),
( 'file', file, (file, mode), 'File')
]:
if obj:
handler = getattr(logging, prefix + 'Handler')(*args)
handler.setFormatter(logging.Formatter(fmt, datefmt))
lo, hi = locals()[name+'_lo_lvl'], locals()[name+'_hi_lvl']
if is_(lo):
handler.setLevel(lo)
if is_(hi):
handler.addFilter(UpToLevel(hi))
logger.addHandler(handler)
if name == 'file':
return handler
def main_module_name(name, ext=True):
if name == '__main__':
try:
main_file = __import__(name).__file__
name_and_ext = main_file[main_file.rfind('/')+1:]
if ext:
return name_and_ext[:name_and_ext.rfind('.')]
return name_and_ext
except:
pass
return name
def get_logger(name, main=False):
name = main_module_name(name) if main else name
return logging.getLogger(ROOT + '.' + name)
### timing
def time_stamp(fmt='%Y-%-m-%-d_%-H-%-M-%-S'):
return dt.datetime.now().strftime(fmt)
### io
def write_lines(iterable, path):
with open(path, 'w') as io:
for item in iterable:
print(item, file=io)
def read_lines(path):
with open(path) as io:
return [line.strip() for line in io]
def mkdir_p(path):
try:
os.makedirs(path)
except OSError as e:
if e.errno != errno.EEXIST:
raise
def get_path_elems_unix(path, i, j='', delim='_'):
elems = re.sub('//+', '/', path).strip('/').split('/')
return elems[i] if j == '' or i == j else delim.join(elems[i:j])
def load(path, method=pickle):
with open(path, 'rb') as f:
return method.load(f)
def dump(obj, path, method=pickle, **kw):
if not kw and method.__name__ in ('pickle', 'dill'):
kw = dict(protocol=-1)
with open(path, 'wb') as f:
method.dump(obj, f, **kw)
def scandir_r(path):
for entry in os.scandir(path):
if entry.is_dir(follow_symlinks=False):
yield from scandir_r(entry.path)
else: yield entry
def mapread(path, f):
with open(path) as io:
yield from map(f, io) if f else io
### convenience
def is_(x):
return x is not None
def dedupe(it):
s = set()
for el in it:
if el not in s:
s.add(el)
yield el
### parsing
def arg(*ar, **kw):
return ar, kw
def strip_attrs(opts, *attrs):
for attr in attrs:
yield getattr(opts, attr)
delattr(opts, attr)
def parse_args(*args, strip=None):
parser = argparse.ArgumentParser()
for ar, kw in args:
parser.add_argument(*ar, **kw)
opts = parser.parse_args()
if is_(strip):
return (opts, *strip_attrs(opts, *strip))
return opts
### sampling
_RNG = random.Random(RANDOM_SEED)
def sample(lst, n=None):
return _RNG.sample(lst, n or len(lst)) | [
"baldsammy@gmail.com"
] | baldsammy@gmail.com |
d50868da0f47f63d568213860fafad9a262e1a9d | 926651d078851f96e71e2cc419fe87348b333ba3 | /mysite/mysite/urls.py | 1e5ccbb2f5ede005dbd394a42c6f06242f90e827 | [] | no_license | Kitabo258/secretary2 | edc9c55133aed40b2d66a89a3d32c90afd923297 | d89e74309abe1577aa4248c77ecc73412e27eba3 | refs/heads/master | 2021-08-14T15:11:43.836979 | 2017-11-16T02:32:12 | 2017-11-16T02:32:12 | 110,913,032 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 907 | py | """mysite URL Configuration
The `urlpatterns` list routes URLs to views. For more information please see:
https://docs.djangoproject.com/en/1.9/topics/http/urls/
Examples:
Function views
1. Add an import: from my_app import views
2. Add a URL to urlpatterns: url(r'^$', views.home, name='home')
Class-based views
1. Add an import: from other_app.views import Home
2. Add a URL to urlpatterns: url(r'^$', Home.as_view(), name='home')
Including another URLconf
1. Import the include() function: from django.conf.urls import url, include
2. Add a URL to urlpatterns: url(r'^blog/', include('blog.urls'))
"""
from django.conf.urls import url
from django.contrib import admin
urlpatterns = [
url(r'^questions/create/$','question.views.question_create'),
url(r'^admin/', admin.site.urls),
url(r'^$', 'show.views.index'),
url(r'^tama/$', 'show.views.tama'),
]
| [
"kiki901028@gmail.com"
] | kiki901028@gmail.com |
1da98ce1969f888ec8962c9239a84d4f7a580f78 | b72dbc51279d3e59cb6410367b671f8a956314c1 | /leet_code/leet_372.py | 5c1d0057a5ac67543ab059922519a69fe52287d6 | [] | no_license | ddobokki/coding-test-practice | 7b16d20403bb1714d97adfd1f47aa7d3ccd7ea4b | c88d981a1d43b986169f7884ff3ef1498e768fc8 | refs/heads/main | 2023-07-08T15:09:32.269059 | 2021-08-08T12:19:44 | 2021-08-08T12:19:44 | 344,116,013 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 205 | py | from typing import List
class Solution:
def superPow(self, a: int, b: List[int]) -> int:
if a in [1,0]:
return a
return int(pow(a,int("".join(str(i) for i in b)),1337))
| [
"44228269+ddobokki@users.noreply.github.com"
] | 44228269+ddobokki@users.noreply.github.com |
d2db1ee6174ddceb75d614d139e0bf52ec6cc9c6 | 65c8e86e276fb9ff8159af867d3793063e480d34 | /config.py | 037c7406057066de211e539d476d582edc9ada7d | [] | no_license | prosass83/ucbotplot | 504d0d16f1625c839d624f95017ca20597b1bf59 | 25b4bb10b4c637d4349ecab56f29a8f9720f7012 | refs/heads/master | 2020-03-10T03:36:40.328166 | 2018-04-12T00:34:43 | 2018-04-12T00:34:43 | 129,169,570 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 270 | py | # Twitter API Keys
consumer_key = "Uu04Kdr6N3SFmbEi334mhy9HQ"
consumer_secret = "d9WJsr5GCCdBI6wbyQopug6HeHpB3T2mKoez8DEHeJFy8D9ko0"
access_token = "2518417008-SC6yochSVIeAYERFWe3BClzJ2pit2iH6YWOUwcA"
access_token_secret = "FBkOuixi1SSXC9xP2k8PpxY07GWsoDZ3MvtC7lj1K59Xo" | [
"prosass@gmail.com"
] | prosass@gmail.com |
8c063266cbd431b2d5053fd81731c057bd7c9d32 | ca1da3b2d522566132ef48319bd2328813a0a0cc | /tests/callbacks/test_progress_bar.py | f621e7022801269b16d12ae7f3d73f8318b9fa95 | [
"Apache-2.0"
] | permissive | anthonytec2/pytorch-lightning | cc4724c3493e93c6c5b89982abe9caf78a969ee2 | 9759491940c4108ac8ef01e0b53b31f03a69b4d6 | refs/heads/master | 2022-11-18T15:48:40.110614 | 2020-07-17T09:54:24 | 2020-07-17T09:54:24 | 273,110,455 | 1 | 0 | Apache-2.0 | 2021-05-13T20:38:51 | 2020-06-18T01:06:34 | Python | UTF-8 | Python | false | false | 6,732 | py | import pytest
from pytorch_lightning import Trainer
from pytorch_lightning.callbacks import ProgressBarBase, ProgressBar, ModelCheckpoint
from pytorch_lightning.utilities.exceptions import MisconfigurationException
from tests.base import EvalModelTemplate
@pytest.mark.parametrize('callbacks,refresh_rate', [
([], 1),
([], 2),
([ProgressBar(refresh_rate=1)], 0),
([ProgressBar(refresh_rate=2)], 0),
([ProgressBar(refresh_rate=2)], 1),
])
def test_progress_bar_on(tmpdir, callbacks, refresh_rate):
"""Test different ways the progress bar can be turned on."""
trainer = Trainer(
default_root_dir=tmpdir,
callbacks=callbacks,
progress_bar_refresh_rate=refresh_rate,
max_epochs=1,
overfit_batches=5,
)
progress_bars = [c for c in trainer.callbacks if isinstance(c, ProgressBarBase)]
# Trainer supports only a single progress bar callback at the moment
assert len(progress_bars) == 1
assert progress_bars[0] is trainer.progress_bar_callback
@pytest.mark.parametrize('callbacks,refresh_rate', [
([], 0),
([], False),
([ModelCheckpoint('../trainer')], 0),
])
def test_progress_bar_off(tmpdir, callbacks, refresh_rate):
"""Test different ways the progress bar can be turned off."""
trainer = Trainer(
default_root_dir=tmpdir,
callbacks=callbacks,
progress_bar_refresh_rate=refresh_rate,
)
progress_bars = [c for c in trainer.callbacks if isinstance(c, ProgressBar)]
assert 0 == len(progress_bars)
assert not trainer.progress_bar_callback
def test_progress_bar_misconfiguration():
"""Test that Trainer doesn't accept multiple progress bars."""
callbacks = [ProgressBar(), ProgressBar(), ModelCheckpoint('../trainer')]
with pytest.raises(MisconfigurationException, match=r'^You added multiple progress bar callbacks'):
Trainer(callbacks=callbacks)
def test_progress_bar_totals(tmpdir):
"""Test that the progress finishes with the correct total steps processed."""
model = EvalModelTemplate()
trainer = Trainer(
default_root_dir=tmpdir,
progress_bar_refresh_rate=1,
limit_val_batches=1.0,
max_epochs=1,
)
bar = trainer.progress_bar_callback
assert 0 == bar.total_train_batches
assert 0 == bar.total_val_batches
assert 0 == bar.total_test_batches
trainer.fit(model)
# check main progress bar total
n = bar.total_train_batches
m = bar.total_val_batches
assert len(trainer.train_dataloader) == n
assert bar.main_progress_bar.total == n + m
# check val progress bar total
assert sum(len(loader) for loader in trainer.val_dataloaders) == m
assert bar.val_progress_bar.total == m
# main progress bar should have reached the end (train batches + val batches)
assert bar.main_progress_bar.n == n + m
assert bar.train_batch_idx == n
# val progress bar should have reached the end
assert bar.val_progress_bar.n == m
assert bar.val_batch_idx == m
# check that the test progress bar is off
assert 0 == bar.total_test_batches
assert bar.test_progress_bar is None
trainer.test(model)
# check test progress bar total
k = bar.total_test_batches
assert sum(len(loader) for loader in trainer.test_dataloaders) == k
assert bar.test_progress_bar.total == k
# test progress bar should have reached the end
assert bar.test_progress_bar.n == k
assert bar.test_batch_idx == k
def test_progress_bar_fast_dev_run(tmpdir):
model = EvalModelTemplate()
trainer = Trainer(
default_root_dir=tmpdir,
fast_dev_run=True,
)
progress_bar = trainer.progress_bar_callback
assert 1 == progress_bar.total_train_batches
# total val batches are known only after val dataloaders have reloaded
trainer.fit(model)
assert 1 == progress_bar.total_val_batches
assert 1 == progress_bar.train_batch_idx
assert 1 == progress_bar.val_batch_idx
assert 0 == progress_bar.test_batch_idx
# the main progress bar should display 2 batches (1 train, 1 val)
assert 2 == progress_bar.main_progress_bar.total
assert 2 == progress_bar.main_progress_bar.n
trainer.test(model)
# the test progress bar should display 1 batch
assert 1 == progress_bar.test_batch_idx
assert 1 == progress_bar.test_progress_bar.total
assert 1 == progress_bar.test_progress_bar.n
@pytest.mark.parametrize('refresh_rate', [0, 1, 50])
def test_progress_bar_progress_refresh(tmpdir, refresh_rate):
"""Test that the three progress bars get correctly updated when using different refresh rates."""
model = EvalModelTemplate()
class CurrentProgressBar(ProgressBar):
train_batches_seen = 0
val_batches_seen = 0
test_batches_seen = 0
def on_batch_start(self, trainer, pl_module):
super().on_batch_start(trainer, pl_module)
assert self.train_batch_idx == trainer.batch_idx
def on_batch_end(self, trainer, pl_module):
super().on_batch_end(trainer, pl_module)
assert self.train_batch_idx == trainer.batch_idx + 1
if not self.is_disabled and self.train_batch_idx % self.refresh_rate == 0:
assert self.main_progress_bar.n == self.train_batch_idx
self.train_batches_seen += 1
def on_validation_batch_end(self, trainer, pl_module):
super().on_validation_batch_end(trainer, pl_module)
if not self.is_disabled and self.val_batch_idx % self.refresh_rate == 0:
assert self.val_progress_bar.n == self.val_batch_idx
self.val_batches_seen += 1
def on_test_batch_end(self, trainer, pl_module):
super().on_test_batch_end(trainer, pl_module)
if not self.is_disabled and self.test_batch_idx % self.refresh_rate == 0:
assert self.test_progress_bar.n == self.test_batch_idx
self.test_batches_seen += 1
progress_bar = CurrentProgressBar(refresh_rate=refresh_rate)
trainer = Trainer(
default_root_dir=tmpdir,
callbacks=[progress_bar],
progress_bar_refresh_rate=101, # should not matter if custom callback provided
limit_train_batches=1.0,
num_sanity_val_steps=2,
max_epochs=3,
)
assert trainer.progress_bar_callback.refresh_rate == refresh_rate
trainer.fit(model)
assert progress_bar.train_batches_seen == 3 * progress_bar.total_train_batches
assert progress_bar.val_batches_seen == 3 * progress_bar.total_val_batches + trainer.num_sanity_val_steps
trainer.test(model)
assert progress_bar.test_batches_seen == progress_bar.total_test_batches
| [
"noreply@github.com"
] | anthonytec2.noreply@github.com |
4f02cd88aa3d26c3be1bbb4b45c2049a6e8a6317 | 9ab9d9a3883471763edbceea59a0e83170581b5f | /eggs/bx_python-0.7.1_7b95ff194725-py2.7-linux-i686-ucs4.egg/EGG-INFO/scripts/bed_extend_to.py | 2985cc3497acf222c69151a76b253624baa01752 | [
"CC-BY-2.5",
"AFL-2.1",
"AFL-3.0",
"CC-BY-3.0",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | asmmhossain/phyG | 24dc211dad5b3e89c87ff384e841f2e98bbd52db | 023f505b705ab953f502cbc55e90612047867583 | refs/heads/master | 2022-11-21T12:43:46.172725 | 2014-02-14T12:33:08 | 2014-02-14T12:33:08 | 13,800,552 | 0 | 1 | NOASSERTION | 2020-07-25T21:05:41 | 2013-10-23T11:04:25 | Python | UTF-8 | Python | false | false | 1,132 | py | #!/afs/bx.psu.edu/project/pythons/linux-i686-ucs4/bin/python2.7
"""
Read BED file and extend each record to the specified minimum length. If chromosome
size information is provided trim extended intervals.
usage: %prog amount [ chrom_file ] < bed_file
"""
import sys
from bx.intervals.io import GenomicIntervalReader
length = int( sys.argv[1] )
chrom_len = None
if len( sys.argv ) > 2:
chrom_len = dict( ( fields[0], int( fields[1] ) ) for fields in map( str.split, open( sys.argv[2] ) ) )
for interval in GenomicIntervalReader( sys.stdin ):
if interval.end - interval.start < length:
start = interval.start
end = interval.end
# Extend in positive direction on strand
if interval.strand == "+":
end = start + length
else:
start = end - length
# Trim
if start < 0:
start = 0
if chrom_len and end > chrom_len[interval.chrom]:
end = chrom_len[interval.chrom]
# Set new start and end
interval.start = start
interval.end = end
# Output possibly adjusted interval
print interval
| [
"mukarram819@gmail.com"
] | mukarram819@gmail.com |
7c3197dc1c975ece63e9829ab29ce34714bbbbbf | 07d5dbbaecddb7ddb3341861fddd5490d54840af | /Computational_Physics/plot2.py | 4da8f397c3aca7afa346f8b38707c1ec770873ec | [] | no_license | julpotter/College-Projects-and-code | 49382e5df82a80c3f2c8162c610c7f465860782e | b1a0508f88a4a6ef1b04de66c9653c68ea03a3cc | refs/heads/main | 2023-03-01T15:25:51.618680 | 2021-02-05T00:54:02 | 2021-02-05T00:54:02 | 326,104,109 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 640 | py |
from __future__ import print
import matplotlib.pyplot as plt
g = 9.81
t0 = 0
y0 = 1.5 # initial height
v0 = float(input("Enter the initial velocity v0 :"))
#use lists to save position and time
y = []
t = []
ycalc = 0
tcalc = 0
#create a loop to populate our lists with values
while(ycalc >= 0):
ycalc = y0 + v0*tcalc - 1/2.*g*tcalc**
y.append(ycalc)
t.append(tcalc)
print "Height is {0:5.2f} and time is {1:5.2f}".format(ycalc,tcalc)
tcalc = tcalc + 0.05 #increment
print "The maximum height reach is ", max(y)
#graph our results
plt.plot(t,y, 'r^')
plt.xlabel("Time t (s)")
plt.ylabel('Height y (m)')
plt.show()
| [
"jpotter@udallas.edu"
] | jpotter@udallas.edu |
f987964ac120b8b838a54bd03b1a6cd553eba48d | e4b52bd79c3a2fdf64f1927ddd04c87fa401fed8 | /myshop/migrations/0003_migrate_translatable_fields.py | cd9bbd763d5d51bc2a494ed80d982b7711caf5e0 | [] | no_license | ObsidianRock/shop | 8e1051efb298c65d6803fe76be07f222c19d60dd | 2a6613f28de3d577288653afc6b1a470fcde6a89 | refs/heads/master | 2021-01-19T10:01:26.797821 | 2017-04-13T16:02:34 | 2017-04-13T16:02:34 | 87,817,222 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,795 | py | # -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
from django.apps import apps
from django.conf import settings
from django.core.exceptions import ObjectDoesNotExist
translatable_models = {
'Category': ['name', 'slug'],
'Product': ['name', 'slug', 'description'],
}
def forwards_func(apps, schema_editor):
for model, fields in translatable_models.items():
Model = apps.get_model('myshop', model)
ModelTranslation = apps.get_model('myshop', '{}Translation'.format(model))
for obj in Model.objects.all():
translation_fields = {field: getattr(obj, field) for field in fields}
translation = ModelTranslation.objects.create(
master_id=obj.pk,
language_code=settings.LANGUAGE_CODE,
**translation_fields)
def backwards_func(apps, schema_editor):
for model, fields in translatable_models.items():
Model = apps.get_model('myshop', model)
ModelTranslation = apps.get_model('myshop', '{}Translation'.format(model))
for obj in Model.objects.all():
translation = _get_translation(obj, ModelTranslation)
for field in fields:
setattr(obj, field, getattr(translation, field))
obj.save()
def _get_translation(obj, MyModelTranslation):
translations = MyModelTranslation.objects.filter(master_id=obj.pk)
try:
return translations.get(language_code=settings.LANGUAGE_CODE)
except ObjectDoesNotExist:
return translations.get()
class Migration(migrations.Migration):
dependencies = [
('myshop', '0002_add_translation_model'),
]
operations = [
migrations.RunPython(forwards_func, backwards_func),
]
| [
"ObsidianRock@users.noreply.github.com"
] | ObsidianRock@users.noreply.github.com |
916795017de2b6b98ecb83425e776e9a8975c090 | 8212c7b8c532681107f735643ab47bb20da177ac | /content_indexer/content_indexer.py | cb4003e22f16b032b2dee9eb2467f20f7d22044f | [] | no_license | jjenner689/Python | 31458c2ca0b612776387fce3bfd3ad56fd07f031 | 665c25077b86b7c802312d5ef4b1ec740b752f8b | refs/heads/master | 2016-08-10T17:06:34.721622 | 2015-10-05T15:00:24 | 2015-10-05T15:00:24 | 43,438,400 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,050 | py | import string
import os
"""
Content Indexing Engine Capstone Project
You have a bunch of files in the file system! How can we index these files to make them easily searchable by keyword?
Indexing is a way of moving work 'upfront' so that when the user searches, less work is needed to get them the right search results.
Tips:
Look into array .extend() method
Look into string module , and .punctuation
Look into the set() builtin data type
Example index:
index = {'cat':['filename1','filename2','filename3'],'dog':['filename2',filename3]}
"""
#Tip: upgrade your recursive find code from a previous exercise to return a list of files
def recursive_find(name, index = {}):
array = os.listdir(name)
for i in array:
path = os.path.join(name, i)
if os.path.isdir(path):
recursive_find(path, index)
else:
data_string = read_data(path)
data_string = strip_punctuation(data_string)
data = split_data_string(data_string)
index = add_to_index(data, i, index)
return index
stop_words = ['a','an','and','i']
def read_data(filename):
with open(filename,"r") as f:
return f.read()
def strip_punctuation(data_string):
punctuation = ["\n",",","'","/","\"","?","+","*","(",")","#","!", "-"]
for i in punctuation:
data_string = data_string.replace(i, ' ')
return data_string
def split_data_string(data_string):
data = data_string.split(" ")
data = list(set(data))
data = map(lambda x: x.lower(), data)
if '' in data:
data.remove('')
return data
def add_to_index(words,filename,index):
for i in words:
if i in index:
index[i].append(filename)
else:
index[i] = [filename]
return index
def handle_words(response, index):
words = response.split(' ')
both = ''
if set(words).issubset(set(index)):
for i in range(len(words)-1):
print index[words[i]], '/', index[words[i+1]]
both = list(set(index[words[i]]) & set(index[words[i+1]]))
if both == '':
both = index[words[0]]
print '\n%s found in files %s' % (words, both)
else:
print '\n%s not found....' % list(set(words) - set(index))
def run_interactive():
print '''\n***Welcome to Josh's content index!***\n'''
#index = recursive_find('/home/josh/Desktop/text_files')
response_1 = ''
response_2 = ''
while not os.path.isdir(response_1):
if response_1 == 'q':
return
response_1 = raw_input('Please enter a valid directory to index or press q to quit > ')
index = recursive_find(response_1)
while response_2 != 'q':
response_2 = raw_input('\nEnter the item/s (separated by spaces) you would like to search or press q to quit > ')
if response_2 == 'q':
break
handle_words(response_2, index)
print "\nThankyou for using Josh's content index.......\n"
run_interactive()
| [
"jjenner689@gmail.com"
] | jjenner689@gmail.com |
52ba6a2814b715ec92f01331d0712c169ef9ebfa | 009eddcbb8e4917710ad3c8661ee5bb04b6fc512 | /tf-transliteration/transliterate.py | 2a032f79be925ab091ef9b896e4cc7c30e5a4c45 | [
"Apache-2.0",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | SRSTLCI/tf-transliteration-riggs2slo-toolkit | 873ea0ccb4c7f270bf16b137ea59ad324ea176ce | e6224627408ea23f830e2b3cd4f6e3a723d70d9d | refs/heads/main | 2023-04-20T18:47:30.731897 | 2021-05-17T16:08:55 | 2021-05-17T16:08:55 | 366,071,013 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,962 | py | from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
import numpy as np
import os, sys, io, re
import six
from data import create_vocab, load_vocab
from data import split_text_file, SPECIALS
from data import create_dataset, make_data_iter_fn
flags = tf.flags
FLAGS = flags.FLAGS
flags.DEFINE_integer("train_steps", 0,
"The number of steps to run training for.")
flags.DEFINE_integer("eval_steps", 100, "Number of steps in evaluation.")
flags.DEFINE_integer("min_eval_frequency", 101, "Minimum steps between evals")
flags.DEFINE_string("hparams", "", "Comma separated list of hyperparameters")
flags.DEFINE_string("model_name", "ei", "Name of model")
flags.DEFINE_string("data_file", None, "TSV Data filename")
flags.DEFINE_float("eval_fraction", 0.05, "Fraction dataset used for evaluation")
flags.DEFINE_string("decode_input_file", None, "File to decode")
flags.DEFINE_string("vocab_file", "chars.vocab", "Character vocabulary file")
tf.logging.set_verbosity(tf.logging.INFO)
def decode_hparams(vocab_size, overrides=""):
hp = tf.contrib.training.HParams(
batch_size=32,
embedding_size=64,
char_vocab_size=vocab_size + 1, #Blank label for CTC loss
hidden_size=128,
learn_rate=0.0008
)
return hp.parse(overrides)
def get_model_dir(model_name):
model_dir = os.path.join(os.getcwd(), model_name)
if not os.path.exists(model_dir):
os.mkdir(model_dir)
return model_dir
def cer(labels, predictions):
dist = tf.edit_distance(predictions, labels)
return tf.metrics.mean(dist)
def create_model():
"""
Actual model function.
Refer https://arxiv.org/abs/1610.09565
"""
def model_fn(features, labels, mode, params):
hparams = params
inputs = features['input']
input_lengths = features['input_length']
targets = labels
target_lengths = features['target_length']
# Flatten input lengths
input_lengths = tf.reshape(input_lengths, [-1])
with tf.device('/cpu:0'):
embeddings = tf.Variable(
tf.truncated_normal(
[hparams.char_vocab_size, hparams.embedding_size],
stddev=(1/np.sqrt(hparams.embedding_size))),
name='embeddings')
input_emb = tf.nn.embedding_lookup(embeddings, inputs)
cell_fw = tf.nn.rnn_cell.BasicLSTMCell(hparams.hidden_size//2)
cell_bw = tf.nn.rnn_cell.BasicLSTMCell(hparams.hidden_size//2)
with tf.variable_scope('encoder'):
# BiLSTM
enc_outputs, _ = tf.nn.bidirectional_dynamic_rnn(cell_fw, cell_bw, input_emb,
input_lengths, dtype=tf.float32)
enc_outputs = tf.concat(enc_outputs, axis=-1)
with tf.variable_scope('decoder'):
# Project to vocab size
logits = tf.layers.dense(enc_outputs, hparams.char_vocab_size)
# CTC loss and decoder requires Time major
logits = tf.transpose(logits, perm=[1, 0, 2])
loss = None
eval_metric_ops = None
train_op = None
predictions = None
if mode == tf.estimator.ModeKeys.TRAIN:
loss = tf.nn.ctc_loss(labels, logits, input_lengths, ignore_longer_outputs_than_inputs=True)
loss = tf.reduce_mean(loss)
optimizer = tf.contrib.opt.LazyAdamOptimizer(learning_rate=hparams.learn_rate)
train_op = optimizer.minimize(loss, global_step=tf.train.get_global_step())
elif mode == tf.estimator.ModeKeys.EVAL:
loss = tf.nn.ctc_loss(labels, logits, input_lengths,
ignore_longer_outputs_than_inputs=True)
loss = tf.reduce_mean(loss)
eval_predictions, _ = tf.nn.ctc_greedy_decoder(logits, input_lengths)
eval_metric_ops = {
'CER': cer(labels, tf.cast(eval_predictions[0], tf.int32))
}
elif mode == tf.estimator.ModeKeys.PREDICT:
predictions, _ = tf.nn.ctc_greedy_decoder(logits, input_lengths)
predictions = tf.sparse_tensor_to_dense(tf.cast(predictions[0], tf.int32))
predictions = {'decoded': predictions}
return tf.estimator.EstimatorSpec(
mode,
predictions=predictions,
loss=loss,
train_op=train_op,
eval_metric_ops=eval_metric_ops
)
return model_fn
def train():
"""
Train the model:
1. Create vocab file from dataset if not created
2. Split dataset into test/eval if not available
3. Create TFRecord files if not available
4. Load TFRecord files using tf.data pipeline
5. Train model using tf.Estimator
"""
model_dir = get_model_dir(FLAGS.model_name)
vocab_file = os.path.join(model_dir, FLAGS.vocab_file)
if not os.path.exists(vocab_file):
create_vocab([FLAGS.data_file], vocab_file)
vocab, characters = load_vocab(vocab_file)
train_file, eval_file = split_text_file(FLAGS.data_file, model_dir, FLAGS.eval_fraction)
train_tfr = create_dataset(train_file, vocab)
eval_tfr = create_dataset(eval_file, vocab)
hparams = decode_hparams(len(vocab), FLAGS.hparams)
tf.logging.info('params: %s', str(hparams))
train_input_fn = make_data_iter_fn(train_tfr, hparams.batch_size, True)
eval_input_fn = make_data_iter_fn(eval_tfr, hparams.batch_size, False)
estimator = tf.estimator.Estimator(
model_fn=create_model(),
model_dir=model_dir,
params=hparams,
config=tf.contrib.learn.RunConfig()
)
experiment = tf.contrib.learn.Experiment(
estimator=estimator,
train_input_fn=train_input_fn,
eval_input_fn=eval_input_fn,
train_steps=FLAGS.train_steps,
eval_steps=FLAGS.eval_steps,
min_eval_frequency=FLAGS.min_eval_frequency
)
experiment.train_and_evaluate()
def predict():
"""
Perform transliteration using trained model. Input must be a text
file. Converts to a TFRecord first.
"""
model_dir = get_model_dir(FLAGS.model_name)
vocab_file = os.path.join(model_dir, FLAGS.vocab_file)
if not os.path.exists(vocab_file):
raise IOError("Could not find vocabulary file")
vocab, rev_vocab = load_vocab(vocab_file)
hparams = decode_hparams(len(vocab), FLAGS.hparams)
tf.logging.info('params: %s', str(hparams))
if FLAGS.decode_input_file is None:
raise ValueError("Must provide input field to decode")
tfr_file = create_dataset(FLAGS.decode_input_file, vocab)
infer_input_fn = make_data_iter_fn(tfr_file, hparams.batch_size, False)
estimator = tf.estimator.Estimator(
model_fn=create_model(),
model_dir=model_dir,
params=hparams,
config=tf.contrib.learn.RunConfig()
)
y = estimator.predict(input_fn=infer_input_fn, predict_keys=['decoded'])
ignore_ids = set([vocab[c] for c in SPECIALS] + [0])
decode_output_file = re.sub(r'\..+', '.out.txt', FLAGS.decode_input_file)
count = 0
with io.open(decode_output_file, 'w', encoding='utf-8') as fp:
for pred in y:
decoded = pred['decoded']
if len(decoded.shape) == 1:
decoded = decoded.reshape(1, -1)
for r in range(decoded.shape[0]):
fp.write(''.join([rev_vocab[i] for i in decoded[r, :] if i not in ignore_ids]) + '\n')
count += 1
if count % 10000 == 0:
tf.logging.info('Decoded %d lines', count)
def main(unused_argv):
if FLAGS.decode_input_file:
predict()
elif FLAGS.train_steps > 0:
train()
tf.app.run()
| [
"noreply@github.com"
] | SRSTLCI.noreply@github.com |
dcb1ace6bff0ebe626977fad8beb5902ed036ab4 | 8a742d701f6e3106f138fb5bb67a90de64c99abd | /lesson1/task2.py | 6dfca1d43eda210d4e9c6cd878da0521bc1d1796 | [] | no_license | alexabolot1/python-homework | ac51f20ad31b866d56783aaffc44d26582f6b113 | 5d719b5f5e3e67ef7e85a8f347d858255cb31ac7 | refs/heads/main | 2023-03-10T16:11:43.824269 | 2021-01-22T06:23:16 | 2021-01-22T06:23:16 | 330,006,488 | 0 | 0 | null | 2021-02-16T12:52:04 | 2021-01-15T19:43:41 | Python | UTF-8 | Python | false | false | 123 | py | a = input('Введите число')
print(f"Сумма вашего числа {int(a) + int(a + a) + int(a + a + a)}")
| [
"alexabolot@gmail.com"
] | alexabolot@gmail.com |
49d7b8bebbad46079eac386c638ee23415db1cdd | 0c8f747e59846ddd7d2479930d38db361ff59bd6 | /secrets.py | b6eb1026c3cc29b7a0e1dd2f39b9bbafa41a8228 | [] | no_license | PaveTranquil/Texity | 39289163e1c5b4ce4d3dea7d718916db8661890a | a4798bbf0cf89496ca12bdb1cd48a3ef9ce4566c | refs/heads/main | 2023-04-13T06:28:32.349938 | 2021-04-27T20:00:52 | 2021-04-27T20:00:52 | 354,387,657 | 3 | 1 | null | 2021-04-06T05:04:37 | 2021-04-03T20:26:48 | Python | UTF-8 | Python | false | false | 69 | py | with open('apikey.txt') as file:
API_KEY = file.readline().strip()
| [
"noreply@github.com"
] | PaveTranquil.noreply@github.com |
59aed57ba62f13ab930efe93f2f9d078ea11c2d3 | f144f0f34227acf229a991d09db18281cd6b5ac6 | /.c9/metadata/environment/abscences/models.py | 269d869f41bb85e539bc1b9b841e06464feed82d | [] | no_license | KikiDow/TUD_HDP_PSE3S | 42dde541a6d65d0c764def842a1e2732fec740ac | cd21c27651e442e1ed84518c1d567f1d52f215c0 | refs/heads/main | 2023-04-22T23:25:25.966531 | 2021-05-09T14:10:01 | 2021-05-09T14:10:01 | 313,276,503 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 817 | py | {"filter":false,"title":"models.py","tooltip":"/abscences/models.py","undoManager":{"mark":2,"position":2,"stack":[[{"start":{"row":8,"column":55},"end":{"row":8,"column":56},"action":"insert","lines":["/"],"id":2},{"start":{"row":8,"column":56},"end":{"row":8,"column":57},"action":"insert","lines":["/"]}],[{"start":{"row":8,"column":56},"end":{"row":8,"column":57},"action":"remove","lines":["/"],"id":3}],[{"start":{"row":8,"column":55},"end":{"row":8,"column":56},"action":"remove","lines":["/"],"id":4}]]},"ace":{"folds":[],"scrolltop":0,"scrollleft":0,"selection":{"start":{"row":8,"column":55},"end":{"row":8,"column":55},"isBackwards":false},"options":{"guessTabSize":true,"useWrapMode":false,"wrapToView":true},"firstLineState":0},"timestamp":1620153966673,"hash":"c83b2801de5e5f3fcb36b6f66f5a7f588f9bceb1"} | [
"ec2-user@ip-172-31-0-153.ec2.internal"
] | ec2-user@ip-172-31-0-153.ec2.internal |
3eeca3eaadcaf592565384f53f3160774272b0ca | 86f5ed0463be0b32508865889f5a77fdda549c22 | /raspberryPi/ultrasonic_client.py | 3441acce60eca5f0c6ba2d110d8c8b66f0352044 | [
"BSD-2-Clause"
] | permissive | pseudoyim/galvaneye | e6d7d210bacab5841e98cf275744dcc4ea6454aa | c7c1f1bb893fb8fb6f83c6765473d506979ec4b0 | refs/heads/master | 2023-01-16T00:43:27.283533 | 2020-11-22T03:09:58 | 2020-11-22T03:09:58 | 70,966,508 | 68 | 25 | null | null | null | null | UTF-8 | Python | false | false | 1,129 | py | from socket import *
import time
import RPi.GPIO as GPIO
GPIO.setwarnings(False)
# create a socket and bind socket to the host
client_socket = socket(AF_INET, SOCK_STREAM)
client_socket.connect(('10.10.10.2', 8002))
def measure():
"""
measure distance
"""
GPIO.output(GPIO_TRIGGER, True)
time.sleep(0.00001)
GPIO.output(GPIO_TRIGGER, False)
start = time.time()
while GPIO.input(GPIO_ECHO)==0:
start = time.time()
while GPIO.input(GPIO_ECHO)==1:
stop = time.time()
elapsed = stop-start
distance = (elapsed * 34300)/2
return distance
# referring to the pins by GPIO numbers
GPIO.setmode(GPIO.BCM)
# define pi GPIO
GPIO_TRIGGER = 16
GPIO_ECHO = 18
# output pin: Trigger
GPIO.setup(GPIO_TRIGGER,GPIO.OUT)
# input pin: Echo
GPIO.setup(GPIO_ECHO,GPIO.IN)
# initialize trigger pin to low
GPIO.output(GPIO_TRIGGER, False)
try:
while True:
distance = measure()
print "Distance : %.1f cm" % distance
# send data to the host every 0.5 sec
client_socket.send(str(distance))
time.sleep(0.5)
finally:
client_socket.close()
GPIO.cleanup()
| [
"paul.j.yim@gmail.com"
] | paul.j.yim@gmail.com |
811cf7cfd7af8000864d8ca0048863f2d418d819 | a9756247d833412e64b18e741c293e7cab9c0b5b | /Command Line/enrollment.py | 3e6491fc442bd5f6d3ef46d8be2a7ca14fa00ac7 | [] | no_license | guennithegun/Command-Line | e44b2b5b7ceac9adc7894f37bf29396b5557dfff | 7c01985129415a56250fa679c9290c3395bf5158 | refs/heads/master | 2020-04-08T04:05:26.323362 | 2018-11-25T05:59:43 | 2018-11-25T05:59:43 | 159,002,688 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 817 | py | import pandas as pd
if __name__ == "__main__":
data = pd.read_csv("data/CRDC2013_14.csv", encoding="Latin-1")
cols =["SCH_ENR_HI_M", "SCH_ENR_HI_F", "SCH_ENR_AM_M", "SCH_ENR_AM_F", "SCH_ENR_AS_M", "SCH_ENR_AS_F", "SCH_ENR_HP_M", "SCH_ENR_HP_F", "SCH_ENR_BL_M", "SCH_ENR_BL_F", "SCH_ENR_WH_M", "SCH_ENR_WH_F", "SCH_ENR_TR_M", "SCH_ENR_TR_F"]
data["total_enrollment"] = data["TOT_ENR_M"] + data["TOT_ENR_F"]
sums = {}
for col in cols:
sums[col] = data[col].sum()
all_enrollment = data["total_enrollment"].sum()
gender_race_perc = {}
for col in cols:
gender_race_perc[col] = (sums[col]*100) / all_enrollment
counter = int()
for keys,values in gender_race_perc.items():
print(keys)
print(values) | [
"noreply@github.com"
] | guennithegun.noreply@github.com |
a03c242a1a98c722662d25054877d5f1d75f61b1 | 8b1ef7f1a7e11b8c39b0546f521a4533f82bb6ba | /PyQt5- Login Form/main.py | 57c90f7324b9bdcd010b805087c425faa78e58d7 | [] | no_license | muhammedzahit/QT5 | 4b6c8e1461f352d4143949486019071285e42d53 | b82283d517023d5860d37e421e5b9babad503469 | refs/heads/master | 2023-02-03T08:19:31.775039 | 2020-12-17T10:25:08 | 2020-12-17T10:25:08 | 289,100,610 | 3 | 0 | null | null | null | null | UTF-8 | Python | false | false | 14,098 | py | # -*- coding: utf-8 -*-
from PyQt5 import QtCore, QtGui, QtWidgets
import sqlite3
class Ui_MainWindow(object):
def setupUi(self, MainWindow):
MainWindow.setObjectName("MainWindow")
MainWindow.resize(800, 600)
sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Fixed)
sizePolicy.setHorizontalStretch(0)
sizePolicy.setVerticalStretch(0)
sizePolicy.setHeightForWidth(MainWindow.sizePolicy().hasHeightForWidth())
MainWindow.setSizePolicy(sizePolicy)
MainWindow.setMinimumSize(QtCore.QSize(800, 600))
MainWindow.setMaximumSize(QtCore.QSize(800, 600))
self.centralwidget = QtWidgets.QWidget(MainWindow)
sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Fixed)
sizePolicy.setHorizontalStretch(0)
sizePolicy.setVerticalStretch(0)
sizePolicy.setHeightForWidth(self.centralwidget.sizePolicy().hasHeightForWidth())
self.centralwidget.setSizePolicy(sizePolicy)
self.centralwidget.setObjectName("centralwidget")
self.bg_photo = QtWidgets.QLabel(self.centralwidget)
self.bg_photo.setGeometry(QtCore.QRect(0, 0, 800, 600))
self.bg_photo.setAutoFillBackground(False)
self.bg_photo.setText("")
self.bg_photo.setPixmap(QtGui.QPixmap("imgs/bg.jpg"))
self.bg_photo.setScaledContents(True)
self.bg_photo.setObjectName("bg_photo")
self.stackedWidget = QtWidgets.QStackedWidget(self.centralwidget)
self.stackedWidget.setGeometry(QtCore.QRect(110, 50, 511, 411))
self.stackedWidget.setObjectName("stackedWidget")
self.page_register = QtWidgets.QWidget()
self.page_register.setStyleSheet("*{\n"
"font: italic 16pt \"Brush Script MT\";\n"
"color : brown;\n"
"background: transparent\n"
"}\n"
"\n"
"QPushButton\n"
"\n"
"{\n"
"background-color : rgb(85, 0, 0, 0.7);\n"
"}\n"
"\n"
"QLabel\n"
"{\n"
"color: yellow\n"
"}\n"
"\n"
"QLineEdit{\n"
"background-color : rgb(85, 0, 0, 0.7);\n"
"color:blue;\n"
"}")
self.page_register.setObjectName("page_register")
self.formLayoutWidget = QtWidgets.QWidget(self.page_register)
self.formLayoutWidget.setGeometry(QtCore.QRect(10, 150, 481, 191))
self.formLayoutWidget.setObjectName("formLayoutWidget")
self.formLayout = QtWidgets.QFormLayout(self.formLayoutWidget)
self.formLayout.setContentsMargins(0, 0, 0, 0)
self.formLayout.setObjectName("formLayout")
self.verticalLayout_4 = QtWidgets.QVBoxLayout()
self.verticalLayout_4.setObjectName("verticalLayout_4")
self.register_name = QtWidgets.QLineEdit(self.formLayoutWidget)
self.register_name.setObjectName("register_name")
self.verticalLayout_4.addWidget(self.register_name)
self.register_password = QtWidgets.QLineEdit(self.formLayoutWidget)
self.register_password.setObjectName("register_password")
self.register_password.setEchoMode(QtWidgets.QLineEdit.Password)
self.verticalLayout_4.addWidget(self.register_password)
self.register_confirm_password = QtWidgets.QLineEdit(self.formLayoutWidget)
self.register_confirm_password.setObjectName("register_confirm_password")
self.register_confirm_password.setEchoMode(QtWidgets.QLineEdit.Password)
self.verticalLayout_4.addWidget(self.register_confirm_password)
self.formLayout.setLayout(0, QtWidgets.QFormLayout.FieldRole, self.verticalLayout_4)
self.verticalLayout_3 = QtWidgets.QVBoxLayout()
self.verticalLayout_3.setObjectName("verticalLayout_3")
self.label_3 = QtWidgets.QLabel(self.formLayoutWidget)
self.label_3.setObjectName("label_3")
self.verticalLayout_3.addWidget(self.label_3)
self.label_4 = QtWidgets.QLabel(self.formLayoutWidget)
self.label_4.setObjectName("label_4")
self.verticalLayout_3.addWidget(self.label_4)
self.label_5 = QtWidgets.QLabel(self.formLayoutWidget)
self.label_5.setObjectName("label_5")
self.verticalLayout_3.addWidget(self.label_5)
self.formLayout.setLayout(0, QtWidgets.QFormLayout.LabelRole, self.verticalLayout_3)
self.registerButton = QtWidgets.QPushButton(self.formLayoutWidget)
self.registerButton.setStyleSheet("")
self.registerButton.setObjectName("registerButton")
self.formLayout.setWidget(1, QtWidgets.QFormLayout.FieldRole, self.registerButton)
self.stackedWidget.addWidget(self.page_register)
self.page_login = QtWidgets.QWidget()
self.page_login.setStyleSheet("*{\n"
"font: italic 16pt \"Brush Script MT\";\n"
"color : brown;\n"
"background: transparent\n"
"}\n"
"\n"
"QPushButton\n"
"\n"
"{\n"
"background-color : rgb(85, 0, 0, 0.7);\n"
"}\n"
"\n"
"QLabel\n"
"{\n"
"color: yellow\n"
"}\n"
"\n"
"QLineEdit{\n"
"background-color : rgb(85, 0, 0, 0.7);\n"
"color:blue;\n"
"}")
self.page_login.setObjectName("page_login")
self.frame = QtWidgets.QFrame(self.page_login)
self.frame.setGeometry(QtCore.QRect(10, 10, 491, 391))
self.frame.setStyleSheet("")
self.frame.setFrameShape(QtWidgets.QFrame.StyledPanel)
self.frame.setFrameShadow(QtWidgets.QFrame.Raised)
self.frame.setObjectName("frame")
self.login_button = QtWidgets.QPushButton(self.frame)
self.login_button.setGeometry(QtCore.QRect(70, 290, 161, 51))
self.login_button.setStyleSheet("")
self.login_button.setObjectName("login_button")
self.to_register_button = QtWidgets.QPushButton(self.frame)
self.to_register_button.setGeometry(QtCore.QRect(260, 290, 161, 51))
self.to_register_button.setStyleSheet("")
self.to_register_button.setObjectName("to_register_button")
self.verticalLayoutWidget = QtWidgets.QWidget(self.frame)
self.verticalLayoutWidget.setGeometry(QtCore.QRect(30, 70, 181, 141))
self.verticalLayoutWidget.setObjectName("verticalLayoutWidget")
self.verticalLayout = QtWidgets.QVBoxLayout(self.verticalLayoutWidget)
self.verticalLayout.setContentsMargins(0, 0, 0, 0)
self.verticalLayout.setObjectName("verticalLayout")
self.label_2 = QtWidgets.QLabel(self.verticalLayoutWidget)
self.label_2.setStyleSheet("")
self.label_2.setAlignment(QtCore.Qt.AlignRight | QtCore.Qt.AlignTrailing | QtCore.Qt.AlignVCenter)
self.label_2.setObjectName("label_2")
self.verticalLayout.addWidget(self.label_2)
self.label = QtWidgets.QLabel(self.verticalLayoutWidget)
self.label.setStyleSheet("")
self.label.setAlignment(QtCore.Qt.AlignRight | QtCore.Qt.AlignTrailing | QtCore.Qt.AlignVCenter)
self.label.setObjectName("label")
self.verticalLayout.addWidget(self.label)
self.verticalLayoutWidget_2 = QtWidgets.QWidget(self.frame)
self.verticalLayoutWidget_2.setGeometry(QtCore.QRect(220, 70, 211, 151))
self.verticalLayoutWidget_2.setObjectName("verticalLayoutWidget_2")
self.verticalLayout_2 = QtWidgets.QVBoxLayout(self.verticalLayoutWidget_2)
self.verticalLayout_2.setContentsMargins(0, 0, 0, 0)
self.verticalLayout_2.setObjectName("verticalLayout_2")
self.login_name = QtWidgets.QLineEdit(self.verticalLayoutWidget_2)
self.login_name.setStyleSheet("")
self.login_name.setObjectName("login_name")
self.verticalLayout_2.addWidget(self.login_name)
self.login_password = QtWidgets.QLineEdit(self.verticalLayoutWidget_2)
self.login_password.setStyleSheet("")
self.login_password.setObjectName("login_password")
self.login_password.setEchoMode(QtWidgets.QLineEdit.Password)
self.verticalLayout_2.addWidget(self.login_password)
self.stackedWidget.addWidget(self.page_login)
MainWindow.setCentralWidget(self.centralwidget)
self.retranslateUi(MainWindow)
self.stackedWidget.setCurrentIndex(1)
QtCore.QMetaObject.connectSlotsByName(MainWindow)
# sayfalar arası geçiş -->
self.to_register_button.clicked.connect(self.to_register_page)
self.registerButton.clicked.connect(self.to_home_page)
self.login_button.clicked.connect(self.check_info)
conn = sqlite3.connect("data.db")
cursor = conn.cursor()
cursor.execute(
"CREATE TABLE IF NOT EXISTS USERS(ID INTEGER NOT NULL PRIMARY KEY, USERNAME TEXT, PASSWORD TEXT)")
conn.close()
def retranslateUi(self, MainWindow):
_translate = QtCore.QCoreApplication.translate
MainWindow.setWindowTitle(_translate("MainWindow", "MainWindow"))
self.label_3.setText(_translate("MainWindow", "Username: "))
self.label_4.setText(_translate("MainWindow", "Password:"))
self.label_5.setText(_translate("MainWindow", "Confirm Password:"))
self.registerButton.setText(_translate("MainWindow", "register"))
self.login_button.setText(_translate("MainWindow", "login"))
self.to_register_button.setText(_translate("MainWindow", "register"))
self.label_2.setText(_translate("MainWindow", "USERNAME :"))
self.label.setText(_translate("MainWindow", "PASSWORD :"))
def to_register_page(self):
self.stackedWidget.setCurrentIndex(0)
def check_username(self, username):
conn = sqlite3.connect("data.db")
curr = conn.cursor()
curr.execute("SELECT USERNAME FROM USERS")
usernames = curr.fetchall()
conn.close()
for name in usernames:
if username == name[0]:
return False
return True
def to_home_page(self):
_translate = QtCore.QCoreApplication.translate
if self.register_name.text() == "" or self.register_password.text() == "" or self.register_confirm_password.text() == "":
msg = QtWidgets.QMessageBox()
msg.setText(_translate("MainWindow", "You must fill all lines !!!"))
msg.setIcon(QtWidgets.QMessageBox.Warning)
x = msg.exec_()
elif self.register_password.text() != self.register_confirm_password.text():
msg = QtWidgets.QMessageBox()
msg.setText(_translate("MainWindow", "Passwords didn't match !!!"))
msg.setIcon(QtWidgets.QMessageBox.Warning)
x = msg.exec_()
elif not self.check_username(self.register_name.text()):
msg = QtWidgets.QMessageBox()
msg.setText(_translate("MainWindow", "Username already taken !!!"))
msg.setIcon(QtWidgets.QMessageBox.Warning)
x = msg.exec_()
else:
conn = sqlite3.connect("data.db")
conn.execute("INSERT INTO USERS(USERNAME,PASSWORD) values(?,?)",
(self.register_name.text(), self.register_password.text(),))
conn.commit()
conn.close()
self.stackedWidget.setCurrentIndex(1)
def check_info(self):
_translate = QtCore.QCoreApplication.translate
conn = sqlite3.connect("data.db")
curr = conn.cursor()
curr.execute("SELECT * FROM USERS")
user_list = curr.fetchall()
print(user_list)
conn.close()
flag = False
for item in user_list:
if item[1] == self.login_name.text() and item[2] == self.login_password.text():
msg = QtWidgets.QMessageBox()
msg.setWindowTitle(_translate("Login", "Logged in"))
msg.setText(_translate("Login", "Login succesfull :)"))
x = msg.exec_()
flag = True
break
if not flag:
msg = QtWidgets.QMessageBox()
msg.setText(_translate("Login", "Username and pasword didn't match !!!"))
msg.setIcon(QtWidgets.QMessageBox.Warning)
msg.setWindowTitle(_translate("Login", "Login Error"))
x = msg.exec_()
else:
exit()
if __name__ == "__main__":
import sys
app = QtWidgets.QApplication(sys.argv)
MainWindow = QtWidgets.QMainWindow()
list = ["Türkçe", "English"]
t = QtCore.QTranslator()
lang = QtWidgets.QInputDialog.getItem(MainWindow, "Select Language", "Language:", list)
#print(lang)
if lang[0] == "Türkçe":
t.load("turkish.qm")
app.installTranslator(t)
ui = Ui_MainWindow()
ui.setupUi(MainWindow)
MainWindow.show()
sys.exit(app.exec_())
| [
"muhammetzahitaydin@gmail.com"
] | muhammetzahitaydin@gmail.com |
4a460011144f616f403fdf0cd4870acf4bd66824 | 3bdc38b3ba7bcd87f10f24fdae3832cc8344ffba | /douban_spider/doubanspider/douban_image_pipelines.py | c34d5d2d875e4668c16b617c5a0411846c223f8e | [] | no_license | sheng-jie/Learning.Python | b1eee7fe53b0b11b8ca4e2e8716dd6ebb06948d6 | 61f0336c76b9a769021238664142cbc0e4a39d02 | refs/heads/master | 2020-03-20T23:01:44.507048 | 2018-06-19T01:46:21 | 2018-06-19T01:46:21 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 856 | py | import os
import scrapy
from scrapy.exceptions import DropItem
from scrapy.pipelines.images import ImagesPipeline
class DoubanImagesPipeline(ImagesPipeline):
def get_media_requests(self, item, info):
for image_url in item['image_urls']:
yield scrapy.Request(image_url, meta={'item': item})
def item_completed(self, results, item, info):
image_paths = [x['path'] for ok, x in results if ok]
if not image_paths:
raise DropItem("Item contains no images")
# os.rename('books/' + image_paths[0], 'books/full/' + item['name'] + '.jpg')
return item
def file_path(self, request, response=None, info=None):
item = request.meta['item']
file_format = request.url.split('.')[-1]
filename = u'full/{0[name]}.{1}'.format(item, file_format)
return filename
| [
"ysjshengjie@live.cn"
] | ysjshengjie@live.cn |
73140cdc70ade106181a0a7092b94bcbb63b6c41 | 4b0ac126af3d635be9d248ed5b2642dfe32b56d0 | /philips_app_engine/main.py | 040715b066a101a64a4565b120c2e2f6f29d8966 | [] | no_license | CosmaTrix/hackathon-git | b4d31a91531818fa943796a81def0a9626283e83 | cfbdf81aa20ff74ce3af3424018f8b90201f5fe4 | refs/heads/master | 2020-12-24T17:08:42.627668 | 2015-01-31T23:08:04 | 2015-01-31T23:08:04 | 30,111,781 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,087 | py | #!/usr/bin/env python
#
# Copyright 2007 Google Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
import time
from urlparse import urljoin
import webapp2
import json
import requests
import settings
GREEN = 21845.
RED = 0
BRIGHTNESS_MAX = 255.
class MainHandler(webapp2.RequestHandler):
def __init__(self, *args, **kwargs):
super(MainHandler, self).__init__(*args, **kwargs)
self.lights = {
0: "http://{0}/api/newdeveloper/lights/1/".format(
settings.PHILIPS_HUE_IP),
1: "http://{0}/api/newdeveloper/lights/3/".format(
settings.PHILIPS_HUE_IP),
2: "http://{0}/api/newdeveloper/lights/2/".format(
settings.PHILIPS_HUE_IP),
}
self.last_light = 2
def __dict_for(self, hue_color, bright):
return {
"on": True,
"sat": 255,
"bri": bright,
"hue": hue_color
}
def __request_dict_from_resp(self, data):
return {
"on": data["state"]["on"],
"sat": data["state"]["sat"],
"br": data["state"]["bri"],
"hue": data["state"]["hue"],
}
def __sequence_lights(self, data):
resp = requests.get(self.lights[1])
data_1 = json.loads(resp.text)
requests.put(urljoin(self.lights[2], 'state'), json.dumps(
self.__request_dict_from_resp(data_1)))
resp = requests.get(self.lights[0])
data_0 = json.loads(resp.text)
requests.put(urljoin(self.lights[1], 'state'), json.dumps(
self.__request_dict_from_resp(data_0)))
requests.put(urljoin(self.lights[0], 'state'), json.dumps(data))
def __turn_lights_off(self):
off_data = json.dumps({"on": False})
requests.put(urljoin(self.lights[0], 'state'), off_data)
time.sleep(0.1)
requests.put(urljoin(self.lights[1], 'state'), off_data)
time.sleep(0.1)
requests.put(urljoin(self.lights[2], 'state'), off_data)
time.sleep(0.1)
def get(self):
fh = open("index.html", "r")
self.response.headers['Content-Type'] = 'text/html'
self.response.out.write(fh.read())
def post(self):
jsonstring = self.request.body
data = json.loads(jsonstring)
values = data.get("values", [])
max_impr = 1
max_vol = 1
tmp_impr = []
tmp_vol = []
for value in values:
impressions = value.get("impressions")
tmp_impr.append(impressions)
volume = value.get("volume")
tmp_vol.append(volume)
max_impr = max(max_impr, impressions)
max_vol = max(max_vol, volume)
rate_impr = GREEN / max_impr
list_impr = [int(impr * rate_impr) for impr in tmp_impr]
rate_vol = BRIGHTNESS_MAX / max_vol
list_vol = [int(vol * rate_vol) for vol in tmp_vol]
response = {}
for i in range(len(list_impr)):
json_dict = self.__dict_for(list_impr[i], list_vol[i])
self.__sequence_lights(json_dict)
json_dict["count"] = i
generated = response.get("generated", [])
generated.append(json_dict)
response["generated"] = generated
time.sleep(data.get("interval", 0.5))
self.__turn_lights_off()
self.response.headers['Content-Type'] = 'application/json'
response["status"] = "OK"
self.response.out.write(json.dumps(response))
app = webapp2.WSGIApplication([('/', MainHandler)], debug=True)
| [
"marco@travelbird.nl"
] | marco@travelbird.nl |
f9d8cfefda12f3f541879b5d40b545e5a08c6842 | 0751fa2615079decfe8c1446f6dcbd7d1048bc31 | /HW3/Code.py | 10d368b3246cb1e5023cca87882b4c37397c8493 | [] | no_license | fatihselimyakar/AlgorithmAndDesign | 0fa4ca8454dc641a57161657ceca29952f45054e | 519107789c77f60dfc03036f7fda962f49203720 | refs/heads/master | 2021-01-06T01:50:07.846213 | 2020-02-17T19:55:35 | 2020-02-17T19:55:35 | 241,194,176 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,280 | py | import itertools
#############QUESTION1###############
def createBoxList( size ):
boxes = []
for i in range(0,(int)(size/2)):
boxes.append("black")
for i in range((int)(size/2),size):
boxes.append("white")
return boxes
def boxTail(boxList,low,middle):
if(middle>=len(boxList)):
return boxList
boxList[low],boxList[middle]=boxList[middle],boxList[low]
boxTail(boxList,low+2,middle+2)
def boxRec(boxList):
return boxTail(boxList,1,(int)(len(boxList)/2))
#############QUESTION2###############
def findFakeCoin( arr ):
if(len(arr)==1):
return arr[0]
if(len(arr)%2==0):
if(sum( arr[0:(int)(len(arr)/2)] ) < sum( arr[(int)(len(arr)/2):len(arr)] )):
return findFakeCoin(arr[0:(int)(len(arr)/2)])
else:
return findFakeCoin(arr[(int)(len(arr)/2):len(arr)])
elif(len(arr)%2==1):
if(sum( arr[0:(int)(len(arr)/2)] ) == sum( arr[(int)(len(arr)/2):len(arr)-1] )):
return arr[len(arr)-1]
elif(sum( arr[0:(int)(len(arr)/2)] ) < sum( arr[(int)(len(arr)/2):len(arr)-1] )):
return findFakeCoin(arr[0:(int)(len(arr)/2)])
else:
return findFakeCoin(arr[(int)(len(arr)/2):len(arr)-1])
#############QUESTION3###############
quicksortSwapNum = 0
insertionSortSwapNum = 0
def insertionSort(arr): #decrease and conquer
global insertionSortSwapNum
for i in range(1,len(arr)):
current=arr[i]
position=i-1
while position>=0 and current<arr[position]:
arr[position+1]=arr[position]
insertionSortSwapNum+=1
position-=1
arr[position+1]=current
return insertionSortSwapNum
def rearrange(arr,low,high):
global quicksortSwapNum
i = ( low-1 )
pivot = arr[high]
for j in range(low , high):
if arr[j] < pivot:
i = i+1
arr[i],arr[j] = arr[j],arr[i]
quicksortSwapNum+=1
arr[i+1],arr[high] = arr[high],arr[i+1]
quicksortSwapNum+=1
return ( i+1 )
def quickSort(arr,low,high):#divide and conquer
if high > low:
index = rearrange(arr,low,high)
quickSort(arr, low, index-1)
quickSort(arr, index+1, high)
return quicksortSwapNum
#############QUESTION4###############
def findMedian(arr):
insertionSort(arr)
if(len(arr)%2==0):
return (arr[(int)(len(arr)/2)]+arr[(int)(len(arr)/2-1)])/2
else:
return arr[(int)(len(arr)/2)]
#############QUESTION5###############
def multiply(numbers):
total = 1
for x in numbers:
total *= x
return total
def optimalSubArray(arr):
value=(max(arr)+min(arr))*(len(arr)/4)
minMult=None
minList=None
for i in range(1,len(arr)+1):
combs=itertools.combinations(arr,i)
minList,minMult = (recSub((list)(combs),value,minList,minMult))
return minList
def recSub(combs,value,minList,minMult):
if(len(combs)==0):
return minList,minMult
elif(sum(combs[0])>=value):
if( (minMult==None) or (multiply(combs[0])<minMult) ):
minMult=multiply(combs[0])
minList=combs[0]
return recSub(combs[1:len(combs)],value,minList,minMult)
def main():
print ("TEST FUNCTION")
print ("\n**Box Test**")
boxList=createBoxList(8)
print ("Unchanged list:",boxList)
boxRec(boxList)
print ("Changed list:",boxList)
print ("\n**Fake Coin Test**")
coins=[2,2,1,2,2,2,2]
print ("Coin list:",coins)
print ("Fake coin:",findFakeCoin(coins))
print ("\n**Insertion and quicksort test**")
arr=[10,9,8,7,6,5,4,3,2,1]
print ("Unsorted array:",arr)
print ("Quicksort number of swap:",quickSort(arr,0,len(arr)-1))
print ("Quicksorted array:",arr)
arr2=[10,9,8,7,6,5,4,3,2,1]
print ("Unsorted array:",arr2)
print ("Insertion sort number of swap:",insertionSort(arr2))
print ("Insertion sorted array:",arr2)
print ("\n**Find median test**")
arr3=[10,20,12,13,19,1]
print ("Median array:",arr3)
print ("Median is:",findMedian(arr3))
print ("\n**Find optimal sub array**")
arr4=[2,4,7,5,22,11]
print ("Array is:",arr4)
print ("Optimal sub array is:",optimalSubArray(arr4))
if __name__ == '__main__':
main() | [
"noreply@github.com"
] | fatihselimyakar.noreply@github.com |
9d58d840d5920137fe056298c28671343069f204 | ee8ee84343e5efd184e20cb474abdd04425aaf7b | /messenger/models.py | 2992db2413316e882741ef7602bad696b1bf617c | [] | no_license | Amrsaeed/codetouch | 978924b59bb68da05f2307fbc2deed31d10aa927 | a70c6d6241e57692501034a5bf6f5d5634738ef1 | refs/heads/master | 2021-01-22T08:18:00.354453 | 2019-01-14T17:05:07 | 2019-01-14T17:05:07 | 92,609,720 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 413 | py | from django.db import models
from django.contrib.auth.models import User
# Create your models here.
class Message(models.Model):
message_text = models.CharField(max_length=2000)
sentOn = models.DateTimeField('Sent On')
sender = models.CharField(max_length=150, default='None')
reciever = models.CharField(max_length=150, default='None')
def __str__(self):
return self.message_text
| [
"amrsaeed@aucegypt.edu"
] | amrsaeed@aucegypt.edu |
795d213349a1ac367a0dcc6c7f13ed3a859b131d | c385ed950cd8512915f97a8bbca466349b647a56 | /code/model.py | c9fc26614280c37bf6aa54a2050dc001f870e95d | [] | no_license | livenb/Ultrasonic_Nerve | b5f2c9c8a0bd0e2e3e654ab247f1dca6fa65b847 | 39f5c036dbf3c63ebaae5c96d86ee6005ccfe3af | refs/heads/master | 2021-01-10T23:08:18.455410 | 2017-02-09T01:42:58 | 2017-02-09T01:42:58 | 70,640,706 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,683 | py | from __future__ import print_function
import cv2
import numpy as np
from keras.models import Model
from keras.layers import Input, merge, Convolution2D, MaxPooling2D, UpSampling2D
from keras.optimizers import Adam
from keras.callbacks import ModelCheckpoint, LearningRateScheduler
from keras import backend as K
from data_prepare import load_train_data, load_test_data
data_path = '../data/'
# K.set_image_dim_ordering('th') # Theano dimension ordering in this code
img_rows = 64
img_cols = 80
smooth = 1.
def dice_coef(y_true, y_pred):
y_true_f = K.flatten(y_true)
y_pred_f = K.flatten(y_pred)
intersection = K.sum(y_true_f * y_pred_f)
return (2. * intersection + smooth) / (K.sum(y_true_f) + K.sum(y_pred_f) + smooth)
def dice_coef_loss(y_true, y_pred):
return -dice_coef(y_true, y_pred)
def get_unet():
inputs = Input((img_rows, img_cols, 1))
conv1 = Convolution2D(32, 3, 3, activation='relu', border_mode='same')(inputs)
conv1 = Convolution2D(32, 3, 3, activation='relu', border_mode='same')(conv1)
pool1 = MaxPooling2D(pool_size=(2, 2))(conv1)
conv2 = Convolution2D(64, 3, 3, activation='relu', border_mode='same')(pool1)
conv2 = Convolution2D(64, 3, 3, activation='relu', border_mode='same')(conv2)
pool2 = MaxPooling2D(pool_size=(2, 2))(conv2)
conv3 = Convolution2D(128, 3, 3, activation='relu', border_mode='same')(pool2)
conv3 = Convolution2D(128, 3, 3, activation='relu', border_mode='same')(conv3)
pool3 = MaxPooling2D(pool_size=(2, 2))(conv3)
conv4 = Convolution2D(256, 3, 3, activation='relu', border_mode='same')(pool3)
conv4 = Convolution2D(256, 3, 3, activation='relu', border_mode='same')(conv4)
pool4 = MaxPooling2D(pool_size=(2, 2))(conv4)
conv5 = Convolution2D(512, 3, 3, activation='relu', border_mode='same')(pool4)
conv5 = Convolution2D(512, 3, 3, activation='relu', border_mode='same')(conv5)
pre = Convolution2D(1, 1, 1, init='he_normal', activation='sigmoid')(conv5)
pre = Flatten()(pre)
aux_out = Dense(1, activation='sigmoid', name='aux_output')(pre)
up6 = merge([UpSampling2D(size=(2, 2))(conv5), conv4], mode='concat', concat_axis=3)
conv6 = Convolution2D(256, 3, 3, activation='relu', border_mode='same')(up6)
conv6 = Convolution2D(256, 3, 3, activation='relu', border_mode='same')(conv6)
up7 = merge([UpSampling2D(size=(2, 2))(conv6), conv3], mode='concat', concat_axis=3)
conv7 = Convolution2D(128, 3, 3, activation='relu', border_mode='same')(up7)
conv7 = Convolution2D(128, 3, 3, activation='relu', border_mode='same')(conv7)
up8 = merge([UpSampling2D(size=(2, 2))(conv7), conv2], mode='concat', concat_axis=3)
conv8 = Convolution2D(64, 3, 3, activation='relu', border_mode='same')(up8)
conv8 = Convolution2D(64, 3, 3, activation='relu', border_mode='same')(conv8)
up9 = merge([UpSampling2D(size=(2, 2))(conv8), conv1], mode='concat', concat_axis=3)
conv9 = Convolution2D(32, 3, 3, activation='relu', border_mode='same')(up9)
conv9 = Convolution2D(32, 3, 3, activation='relu', border_mode='same')(conv9)
conv10 = Convolution2D(1, 1, 1, activation='sigmoid',
name='main_output')(conv9)
model = Model(input=inputs, output=[conv10, aux_out])
model.compile(optimizer=Adam(lr=1e-5),
loss={'main_output': dice_coef_loss, 'aux_output': 'binary_crossentropy'},
metrics={'main_output': dice_coef, 'aux_output': 'acc'},
loss_weights={'main_output': 1., 'aux_output': 0.5})
return model
def preprocess(imgs):
imgs_p = np.ndarray((imgs.shape[0], img_rows, img_cols, 1),
dtype=np.uint8)
for i in range(imgs.shape[0]):
img = cv2.resize(imgs[i], (img_cols, img_rows))
imgs_p[i] = img.reshape((img.shape[0],img.shape[1],1))
return imgs_p
def mask_exist(mask):
return np.array([int(np.sum(mask[i, 0]) > 0) for i in xrange(len(mask))])
def train_and_predict():
print('-'*30)
print('Loading and preprocessing train data...')
print('-'*30)
imgs_train, imgs_mask_train = load_train_data()
imgs_train = preprocess(imgs_train)
imgs_mask_train = preprocess(imgs_mask_train)
imgs_train = imgs_train.astype('float32')
mean = np.mean(imgs_train) # mean for data centering
std = np.std(imgs_train) # std for data normalization
imgs_train -= mean
imgs_train /= std
imgs_mask_train = imgs_mask_train.astype('float32')
imgs_mask_train /= 255. # scale masks to [0, 1]
print('-'*30)
print('Creating and compiling model...')
print('-'*30)
model = get_unet()
model_checkpoint = ModelCheckpoint(data_path+'unet.hdf5', monitor='loss',
save_best_only=True)
print('-'*30)
print('Fitting model...')
print('-'*30)
model.fit(imgs_train, [imgs_mask_train, mask_exist(imgs_mask_train)],
batch_size=32, nb_epoch=20,
verbose=1, shuffle=True, callbacks=[model_checkpoint])
print('-'*30)
print('Loading and preprocessing test data...')
print('-'*30)
imgs_test, imgs_id_test = load_test_data()
imgs_test = preprocess(imgs_test)
imgs_test = imgs_test.astype('float32')
imgs_test -= mean
imgs_test /= std
print('-'*30)
print('Loading saved weights...')
print('-'*30)
model.load_weights(data_path+'unet.hdf5')
print('-'*30)
print('Predicting masks on test data...')
print('-'*30)
imgs_mask_test = model.predict(imgs_test, verbose=1)
np.save(data_path+'imgs_mask_test.npy', imgs_mask_test)
if __name__ == '__main__':
train_and_predict()
| [
"livenb666@gmail.com"
] | livenb666@gmail.com |
c8df0b5f2035c2386b9918776f917c3effb9da50 | 02fa1542bc428b64da276afdb46f2f2f7199f7a6 | /DataManager.py | ff0a2123a51d8a90d13e0f767a8e98c9f0b921d1 | [] | no_license | ulissesbcorrea/atae-lstm-theano-modified | bb4ba7b9786d00d5bf6da16cbbdd5067e8118e91 | e3bfa6a74df878ee7eec115c1a6e8fd1c4fcfa46 | refs/heads/master | 2022-12-13T16:36:10.256232 | 2020-09-14T09:21:14 | 2020-09-14T09:21:14 | 287,218,745 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,457 | py | # -*- encoding: utf-8 -*-
import numpy as np
import theano
class Sentence(object):
"""docstring for sentence"""
def __init__(self, content, target, rating, grained):
self.content, self.target = content.lower(), target
self.solution = np.zeros(grained, dtype=theano.config.floatX)
self.senlength = len(self.content.split(' '))
try:
self.solution[int(rating)+1] = 1
except Exception as e:
print 'erro no contrutor de Sentence:'+ str(e)
print 'rating:' + rating
exit()
def stat(self, target_dict, wordlist, grained=3):
data, data_target, i = [], [], 0
solution = np.zeros((self.senlength, grained), dtype=theano.config.floatX)
for word in self.content.split(' '):
data.append(wordlist[word])
# try:
# pol = Lexicons_dict[word]
# solution[i][pol+1] = 1
# except Exception as e:
# print 'error in stat:' + str(e)
# pass
i = i+1
for word in self.target.split(' '):
data_target.append(wordlist[word])
return {'seqs': data, 'target': data_target, 'solution': np.array([self.solution]), 'target_index': self.get_target(target_dict), 'original_text':self.content, 'aspect': self.target}
def get_target(self, dict_target):
return dict_target[self.target]
class DataManager(object):
def __init__(self, dataset, seed, grained=3):
self.fileList = ['train', 'test', 'dev']
self.origin = {}
for fname in self.fileList:
data = []
with open('%s/%s.cor' % (dataset, fname)) as f:
sentences = f.readlines()
for i in xrange(len(sentences)/3):
content, target, rating = sentences[i*3].strip(), sentences[i*3+1].strip(), sentences[i*3+2].strip()
sentence = Sentence(content, target, rating, grained)
data.append(sentence)
self.origin[fname] = data
self.gen_target()
def gen_word(self):
wordcount = {}
def sta(sentence):
for word in sentence.content.split(' '):
try:
wordcount[word] = wordcount.get(word, 0) + 1
except:
wordcount[word] = 1
for word in sentence.target.split(' '):
try:
wordcount[word] = wordcount.get(word, 0) + 1
except:
wordcount[word] = 1
for fname in self.fileList:
for sent in self.origin[fname]:
sta(sent)
words = wordcount.items()
words.sort(key=lambda x:x[1], reverse=True)
self.wordlist = {item[0]:index+1 for index, item in enumerate(words)}
return self.wordlist
def gen_target(self, threshold=5):
self.dict_target = {}
for fname in self.fileList:
for sent in self.origin[fname]:
if self.dict_target.has_key(sent.target):
self.dict_target[sent.target] = self.dict_target[sent.target] + 1
else:
self.dict_target[sent.target] = 1
i = 0
for (key,val) in self.dict_target.items():
if val < threshold:
self.dict_target[key] = 0
else:
self.dict_target[key] = i
i = i + 1
return self.dict_target
def gen_data(self, grained=3):
self.data = {}
for fname in self.fileList:
self.data[fname] = []
for sent in self.origin[fname]:
self.data[fname].append(sent.stat(self.dict_target, self.wordlist))
return self.data['train'], self.data['dev'], self.data['test']
def word2vec_pre_select(self, mdict, word2vec_file_path, save_vec_file_path):
list_seledted = ['']
line = ''
with open(word2vec_file_path) as f:
for line in f:
tmp = line.strip().split(' ', 1)
if mdict.has_key(tmp[0]):
list_seledted.append(line.strip())
list_seledted[0] = str(len(list_seledted)-1) + ' ' + str(len(line.strip().split())-1)
open(save_vec_file_path, 'w').write('\n'.join(list_seledted))
| [
"ulissesbcorrea@gmail.com"
] | ulissesbcorrea@gmail.com |
cccee8c95ce17bb44043b1a20a899ac4161055be | ee22ec2076a79e8de3011377fe205bc87163ab9f | /src/basic-c3/func-let.py | 8c9c6ff3fea14adfbe60b86692ad4981a5710241 | [] | no_license | n18018/programming-term2 | 039a95c67372a38a34e2aa8c5975045a9fc731be | 86c455269eed312def529604e1ac3b00f476226c | refs/heads/master | 2020-03-22T08:59:29.545280 | 2018-08-29T07:57:37 | 2018-08-29T07:57:37 | 139,806,131 | 0 | 0 | null | 2018-07-05T06:42:11 | 2018-07-05T06:42:11 | null | UTF-8 | Python | false | false | 326 | py | # 関数を定義
def mul_func(a, b):
return a * b
def div_func(a, b):
return a / b
# mul_func関数を変数に代入
func = mul_func
# 代入した変数で関数を使う
result = func(2, 3)
print(result)
# div_func関数を変数に代入する場合
func2 = div_func
result = func2(10, 5)
print(result)
| [
"n18018@std.it-college.ac.jp"
] | n18018@std.it-college.ac.jp |
614d20e490badf198728acb806f1b442ff8a43b7 | 9ab642dbc8b5409673e5b2f90e009aa4b5634c32 | /st_network_server.py | ea0366083d46d6693412626cfecd9193cb995a67 | [] | no_license | sunpu/stwhiteboard | 65c20aab6049acc0ca1b5b1924ae048671e119ee | 26cb02b593076e65496beca162be24404bee690a | refs/heads/master | 2021-09-07T12:40:49.138494 | 2018-02-23T01:44:37 | 2018-02-23T01:44:37 | 115,606,494 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,122 | py | #!/usr/bin/python2.7
# -*- coding:utf-8 -*-
import SocketServer
from time import ctime
import json
courseDict = {}
class Course:
def __init__(self, courseID):
self.courseID = courseID
self.clientDict = {}
self.finishList = []
HOST = ''
PORT = 10001
ADDR = (HOST, PORT)
ROOT_PATH = './file/'
class Client(SocketServer.BaseRequestHandler):
role = ''
courseID = 0
bigData = ''
def readDirectory(self, path):
result = []
paths = os.listdir(path)
for i, item in enumerate(paths):
sub_path = os.path.join(path, item)
data = {}
data['name'] = item
timestamp = os.path.getmtime(sub_path)
date = datetime.datetime.fromtimestamp(timestamp)
data['time'] = date.strftime('%Y-%m-%d %H:%M:%S')
if os.path.isdir(sub_path):
data['type'] = 'folder'
data['size'] = '-'
else:
data['type'] = 'file'
fsize = os.path.getsize(sub_path)
fsize = fsize / float(1024)
data['size'] = str(round(fsize,2)) + 'KB'
result.append(data)
json_res = json.dumps(result)
return json_res
def sendMessage(self, data):
self.request.sendall('#*#' + data + '@%@')
def sendHistoryMessage(self):
#print courseDict[self.courseID].finishList
finishList = courseDict[self.courseID].finishList
for index in range(0, len(finishList)):
self.request.sendall('#*#' + finishList[index] + '@%@')
def boardcastMessage(self, data):
#print courseDict
clientDict = courseDict[self.courseID].clientDict
#print clientDict
for k in clientDict.keys():
if k == self.client_address:
continue
#print k, clientDict[k]
#print '---', data
clientDict[k].sendall('#*#' + data + '@%@')
def processData(self, data):
#print '--------------', data
datas = json.loads(data)
if datas['type'] == 'createClient':
self.role = datas['data']['role']
elif datas['type'] == 'createCourse':
self.courseID = datas['data']['courseID']
if courseDict.has_key(self.courseID):
return
course = Course(self.courseID)
courseDict[self.courseID] = course
elif datas['type'] == 'joinCourse':
self.courseID = datas['data']['courseID']
course = courseDict[self.courseID]
course.clientDict[self.client_address] = self.request
self.sendHistoryMessage()
elif datas['type'] == 'setClientAuthority' or datas['type'] == 'finish':
self.boardcastMessage(data)
course = courseDict[self.courseID]
for index in range(0, len(course.finishList)):
#print '---', course.finishList[index]
historyDatas = json.loads(course.finishList[index])
#print 'historyDatas---', historyDatas
#print 'datas---', datas
if historyDatas.has_key('itemID') and datas.has_key('itemID') and historyDatas['itemID'] == datas['itemID'] and historyDatas['subtype'] == datas['subtype']:
course.finishList.remove(course.finishList[index])
break
course.finishList.append(data)
elif datas['type'] == 'realtime':
self.boardcastMessage(data)
elif datas['type'] == 'file':
path = ROOT_PATH
if datas['action'] == 'list':
path += datas['data']['path']
elif datas['action'] == 'new':
path += datas['data']['path']
name = datas['data']['name']
cmd = 'cd %s;mkdir %s;' % (path, name)
os.system(cmd)
elif datas['action'] == 'copy':
path += datas['data']['path']
name = datas['data']['name']
destPath += datas['data']['destPath']
cmd = 'cd %s;cp -rf %s %s;' % (path, name, destPath)
os.system(cmd)
elif datas['action'] == 'move':
path += datas['data']['path']
name = datas['data']['name']
destPath += datas['data']['destPath']
cmd = 'cd %s;mv -rf %s %s;' % (path, name, destPath)
os.system(cmd)
elif datas['action'] == 'del':
path += datas['data']['path']
name = datas['data']['name']
cmd = 'cd %s;rm -rf %s;' % (path, name)
os.system(cmd)
list = self.readDirectory(path)
self.sendMessage(list)
#{"type":"file","action":"list","data":{"path":"/1/2"}}
#{"type":"file","action":"new","data":{"path":"/1/2","name":"xxx"}}
#{"type":"file","action":"copy","data":{"path":"/2","name":"xxx","destPath":"/3"}}
#{"type":"file","action":"move","data":{"path":"/2","name":"xxx","destPath":"/3"}}
#{"type":"file","action":"del","data":{"path":"/2","name":"xxx"}}
def handle(self):
# 客户端登入后,记住
print '...connected from:', self.client_address
while True:
data = self.request.recv(1024 * 1024 * 10)
#print data, 'from', self.client_address
if len(data) == 0:
course = courseDict[self.courseID]
course.clientDict.pop(self.client_address)
break
if data.endswith('@%@'):
if len(self.bigData) > 0:
data = self.bigData + data
self.bigData = ''
data = data.replace('@%@', '')
dataList = data.split('#*#')
for index in range(0, len(dataList)):
if dataList[index]:
self.processData(dataList[index])
else:
self.bigData = self.bigData + data
tcpServ = SocketServer.ThreadingTCPServer(ADDR, Client)
print 'waiting for connection...'
tcpServ.serve_forever() | [
"sunpumsn@hotmail.com"
] | sunpumsn@hotmail.com |
820e15a0e02606bf4165dcedc96fab3b641663ef | da8a2b9404e6bb9f3d6ca5a786fd01eddf440ec4 | /lyric/apps.py | 618476d0dd10601771f2dbee5c71aa66f38b89f0 | [] | no_license | BrightHao/Django_Music | c099e22d4606a8101f1be3be3a9f4e7b2babcc25 | c31df1a0d3f34594c65bddcc88815e22bc2f0902 | refs/heads/master | 2023-03-17T20:01:33.915667 | 2021-03-13T05:37:49 | 2021-03-13T05:37:49 | 347,283,264 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 85 | py | from django.apps import AppConfig
class LyricConfig(AppConfig):
name = 'lyric'
| [
"861759757@qq.com"
] | 861759757@qq.com |
ae8bf909464124ce2e2e1c318f37dd319d3ef4ac | 5a99d1f7e0363878a5a94732598410a06008d2ed | /multimeter/_tasks.py | 204079b61c271657c9be403e1e4f4d875d9109f1 | [] | no_license | av-pavlov/multimeter | d8b93e4a6acec420e8593ea8864db9d67ab4a177 | d8132db1e5e0c3b153ab142599c93b302c18fbf5 | refs/heads/master | 2021-04-29T18:44:15.300083 | 2018-03-23T04:38:45 | 2018-03-23T04:38:45 | 121,699,767 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 12,491 | py | # -*- coding: utf-8 -*=
import sys
import subprocess
from collections import OrderedDict
from os import listdir, stat
from os.path import isdir, join, isfile
from collections import OrderedDict
from .helpers import load_json, save_json, validate_code, check_or_create_dir, load_tests
class Tasks:
""" Массив олимпиадных задач """
def __init__(self, settings, languages):
self._settings = settings
self._languages = languages
self.tasks = OrderedDict()
self.load()
def __len__(self):
return len(self.tasks)
def __setitem__(self, key, value):
self.tasks[key] = value
def __getitem__(self, item):
return self.tasks[item]
def __delitem__(self, key):
del self.tasks[key]
def __contains__(self, item):
return item in self.tasks
def __iter__(self):
return self.tasks.__iter__()
def items(self):
return self.tasks.items()
def keys(self):
return sorted(self.tasks.keys())
def load(self):
""" Загрузить олимпиадные задачи из подкаталогов рабочего каталога """
# Просмотрим подкаталоги рабочего каталога
dirs = sorted(listdir(self._settings.work_dir))
for name in dirs:
path = join(self._settings.work_dir, name)
if isdir(path) and '.' not in path:
try:
self.tasks[name] = Task(name, path)
self.tasks[name].load()
except (TypeError, FileNotFoundError, UnicodeDecodeError):
# Если не удалось прочитать описание задачи - игнорируем этот подкаталог
pass
def save(self):
""" Сохранить описания олимпиадных задач в их подкаталогах """
for task in self.tasks:
task.save()
def get_results(self, task_code, username, attempt=None):
""" Получить результаты проверки решений олимпиадной задачи определенным пользователем """
answer = []
# Начнем просмотр файлов результатов в каталоге .results
results_dir = join(self._settings.work_dir, '.results')
for filename in listdir(results_dir):
# Только только JSON-файлы
if filename[-5:] != '.json':
continue
name = filename[:-5]
(_task_code, _username, _attempt) = name.split('-')
# Только выбранная олимпиадная задача
if task_code != _task_code:
continue
# Только задачи определенного пользователя
if username != _username:
continue
# Если была выбрана попытка, то нужна только определенная попытка
if attempt is not None and attempt != _attempt:
continue
# Прочитаем результат проверки и добавим в номер попытки
res = load_json(filename, {}, results_dir)
res['attempt'] = int(_attempt)
answer.append(res)
return sorted(answer, key=lambda x: x['attempt'])
def validate_task(self, code, data, check_uniqueness):
""" Проверка задания
:param check_uniqueness:
:param data:
:param code:
"""
codes_list = self.tasks if check_uniqueness else []
errors = validate_code(code, codes_list, 'Код задания')
if not data.get('name'):
errors.append('Наименование не задано')
return errors
class TestSuite:
# Стратегия отображения результатов
BRIEF = 'brief'
FULL = 'full'
ERROR = 'error'
RESULTS = (
(BRIEF, 'Отображаются только баллы за подзадачу целом'),
(FULL, 'Отображаются баллы за каждый тест'),
(ERROR, 'Отображаются баллы за подзадачу в целом либо результат первой ошибки'),
)
# Стратегия начисления баллов
PARTIAL = 'partial'
ENTIRE = 'entire'
SCORING = (
(PARTIAL, 'Баллы начисляются пропорционально'),
(ENTIRE, 'Подзадача оценивается как единое целое'),
)
task = None
code = ''
ts_dir = ''
name = ''
results = FULL
scoring = PARTIAL
test_score = 0
total_score = 0
depends = []
def __init__(self, task, code, data):
self.task = task
self.code = code
self.ts_dir = join(task.test_suites_dir, code)
self.name = data['name']
self.scoring = data['scoring']
self.results = data['results']
self.test_score = data.get('test_score', 0)
self.total_score = data.get('total_score', 0)
self.tests = load_tests(self.ts_dir)
self.depends = data.get('depends', self.depends)
class Task:
# Важные атрибуты
code = '' # Код
task_dir = '' # Каталог
# Атрибуты из конфигурационного файла
name = '' # Имя
timeout = 2.0 # Предельное время выполнения в секундах, при превышении - работа программа будет завершена
time_limit = 1.0 # Лимит времени выполнения в секундах, при превышении - вердикт TL
memory_limit = 256 # Лимит по количеству памяти в Мб, при превышении - вердикт ML
input_file = 'input.txt' # Имя выходного файла
output_file = 'output.txt' # Имя выходного файла
# Атрибуты заполняемые из файлов
statement = '' # Условия задачи
preliminary = [] # Список примеров для предварительной проверки решения
test_suites = OrderedDict() # Словарь подзадач, подзадача - это список тестов
def __init__(self, code, task_dir):
"""
Создание задачи по каталогу
:param code: код задачи
:param task_dir: каталог задачи
"""
self.code = code
self.task_dir = task_dir
@property
def brief_name(self):
return '%s. %s' % (self.code, self.name)
@property
def full_name(self):
return 'Задача %s. %s' % (self.code, self.name)
@property
def config_file(self):
return join(self.task_dir, 'task.json')
@property
def statements_file(self):
return join(self.task_dir, 'task.html')
@property
def checker(self):
return join(self.task_dir, 'check.exe')
@property
def solutions_dir(self):
return join(self.task_dir, 'solutions')
@property
def preliminary_dir(self):
return join(self.task_dir, 'tests', 'samples')
@property
def test_suites_dir(self):
return join(self.task_dir, 'tests')
def load(self):
""" Читаем описание задачи из конфигурационных файлов """
# Загружаем атрибуты задачи из конфигурационного файла
config = load_json(self.config_file, {})
if 'name' in config:
self.name = str(config['name'])
if 'timeout' in config:
self.timeout = float(config['timeout'])
if 'time_limit' in config:
self.time_limit = float(config['time_limit'])
if 'memory_limit' in config:
self.memory_limit = float(config['memory_limit'])
if 'input_file' in config:
self.input_file = str(config['input_file'])
if 'output_file' in config:
self.output_file = str(config['output_file'])
if 'test_suites' in config:
tss_from_file = config['test_suites']
if isinstance(tss_from_file, OrderedDict):
self.test_suites = OrderedDict()
for code, ts in tss_from_file.items():
self.test_suites[code] = TestSuite(self, code, ts)
# Загружаем условия задачи
try:
statement = open(self.statements_file, encoding='utf-8')
self.statement = statement.read()
except FileNotFoundError:
# Если файла нет - молча ничего не делаем
pass
# Загружаем примеры
self.preliminary = load_tests(self.preliminary_dir)
def save(self):
""" Сохранение задачи в task.json в каталоге задачи """
keys = ['name', 'brief_name', 'timeout', 'input_file', 'output_file', 'test_suites']
config = dict(zip(keys, [self.__dict__[k] for k in keys]))
save_json(config, self.config_file)
with open(self.statements_file, mode='w', encoding='utf-8') as f:
f.write(self.statement)
f.close()
def verify(self):
if not isdir(self.task_dir):
raise Exception('Task {} folder not found: {}'.format(self.code, self.task_dir))
check_or_create_dir(self.solutions_dir)
check_or_create_dir(self.test_suites_dir)
for test in self.preliminary:
self.verify_test(test)
total_score = 0
for suite_code, suite in self.test_suites.items():
if suite.scoring == TestSuite.ENTIRE:
total_score += suite.total_score
elif suite.scoring == TestSuite.PARTIAL:
total_score += suite.test_score * len(suite.tests)
for test in suite.tests:
self.verify_test(test, suite_code)
if total_score != 100:
raise Exception('Sum of tests score of task {} not equal 100 !!!'.format(self.code))
def verify_test(self, test, suite_code=None):
""" Проверка теста
:param test: имя теста
:param suite_code:
"""
if suite_code is None:
test_name = "Preliminary test {}".format(test)
input_file = join(self.preliminary_dir, test)
else:
test_name = "Test {} in {}".format(test, suite_code)
input_file = join(self.test_suites_dir, suite_code, test)
answer_file = input_file + '.a'
if not isfile(input_file):
raise Exception('{} for task {} not found !!!'.format(test_name, self.code))
if not isfile(answer_file):
raise Exception('{} for task {} don\'t have answer !!!'.format(test_name, self.code))
try:
subprocess.check_call(
[self.checker, input_file, answer_file, answer_file],
stderr=subprocess.DEVNULL,
stdout=subprocess.DEVNULL)
except FileNotFoundError:
raise Exception('Checker for task {} is not found !!!'.format(self.code))
except subprocess.CalledProcessError:
raise Exception('Checker for task {} is not working !!!'.format(self.code))
def check(self):
""" Проверка ответа участника """
answer = '??'
try:
output_file = 'stdout'
if isfile(self.output_file) and stat(self.output_file).st_size > 0:
output_file = self.output_file
subprocess.check_call([
self.checker,
self.input_file,
output_file,
'answer.txt',
])
answer = 'OK'
except subprocess.CalledProcessError as error:
if error.returncode == 1:
answer = 'WA' # Wrong answer
elif error.returncode == 2:
answer = 'PE' # Presentation error
finally:
return answer
| [
"7911562+av-pavlov@users.noreply.github.com"
] | 7911562+av-pavlov@users.noreply.github.com |
b04bcc13154c9db03c343251a42fa7662c29b486 | 52c737c2f062f7b07e35ab88e9ea042337463253 | /Weibo_v3/weibo_auto_handle.py | b1c5aade7198eb39ae4667657b948245d1f5a156 | [] | no_license | LichMscy/Weibo | b704779cdfe8c52063bb3c8e2d4e611ffd47bb37 | 6d3ed1e9709ef48c4627dfc69fd9ee9cbe7e628b | refs/heads/master | 2021-01-18T17:26:20.351718 | 2017-04-01T02:47:49 | 2017-04-01T02:47:49 | 86,798,194 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 7,101 | py | # -*-coding: utf-8 -*-
import re
import time
import datetime
import logging
import json
from selenium import webdriver
from selenium.webdriver.common.desired_capabilities import DesiredCapabilities
from selenium.webdriver.common.keys import Keys
import requests
from bs4 import BeautifulSoup
from util import parsedate
import persist_iics
logging.basicConfig(level=logging.INFO, format="[%(asctime)s]%(name)s:%(levelname)s:%(message)s")
logger = logging.getLogger(__name__)
logging.getLogger("selenium").setLevel(logging.WARNING)
def init_phantomjs_driver():
headers = {
'Cookie': 'YF-Ugrow-G0=b02489d329584fca03ad6347fc915997; SUB=_2AkMvgPj2dcPxrAFYnPgWyGvkZYpH-jycVZEAAn7uJhMyOhgv7nBSqSVOKynW2PbhU4768kfRGZgNPwXeRA..; SUBP=0033WrSXqPxfM72wWs9jqgMF55529P9D9WWEFXHsNpvgJdQjr1GM.e765JpVF020SKM7e0571hMc',
}
for key, value in headers.items():
webdriver.DesiredCapabilities.PHANTOMJS['phantomjs.page.customHeaders.{}'.format(key)] = value
useragent = 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/57.0.2987.110 Safari/537.36'
webdriver.DesiredCapabilities.PHANTOMJS['phantomjs.page.settings.userAgent'] = useragent
# local path refer phantomjs
driver = webdriver.PhantomJS(executable_path='xxxxx')
driver.set_window_size(1366, 768)
return driver
def update_cookies():
p1 = persist_iics.Persist()
accounts = p1.query_account()
cookie = json.loads(accounts[0][3])
req = requests.Session().get('http://weibo.cn/', cookies=cookie)
if re.findall('登录|注册', req.text, re.S):
logging.error('Account cookies out of date! (Account_id: %s)' % accounts[0][0])
browser = init_phantomjs_driver()
try:
browser.get("http://weibo.com")
time.sleep(3)
failure = 0
while "微博-随时随地发现新鲜事" == browser.title and failure < 5:
failure += 1
username = browser.find_element_by_name("username")
pwd = browser.find_element_by_name("password")
login_submit = browser.find_element_by_class_name('W_btn_a')
username.clear()
username.send_keys(accounts[0][1])
pwd.clear()
pwd.send_keys(accounts[0][2])
login_submit.click()
time.sleep(5)
# if browser.find_element_by_class_name('verify').is_displayed():
# logger.error("Verify code is needed! (Account: %s)" % account)
if "我的首页 微博-随时随地发现新鲜事" in browser.title:
browser.get('http://weibo.cn/')
cookies = dict()
if "我的首页" in browser.title:
for elem in browser.get_cookies():
cookies[elem["name"]] = elem["value"]
p2 = persist_iics.Persist()
p2.save_account_cookies(accounts[0][0], cookies, datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S"))
logging.error('Account cookies updated! (Account_id: %s)' % accounts[0][0])
return cookies
except:
logger.error("Weibo Login Unknown exception!")
raise
else:
return cookie
def snatch_news_info(cookies):
p1 = persist_iics.Persist()
p1_result = p1.query_task()
ids = re.findall('weibo.com/(.*?)/(.*?)[?]', p1_result[0][1], re.S)[0]
if ids and ids[0] and ids[1]:
url = 'http://weibo.cn/comment/{}?uid={}'.format(ids[1], ids[0])
req = requests.get(url, cookies=cookies)
while req.status_code != 200:
logging.error('Snatch (Task_id: %s) failed!' % p1_result[0][0])
exit()
soup = BeautifulSoup(req.text, 'lxml')
item = soup.select('span.ctt')[0]
dic = dict()
dic['platform_id'] = 2
dic['media_name'] = '新浪微博'
dic['title'] = item.get_text()[1:23] + '...'
dic['summary'] = item.get_text()[1:290]
dic['src_url'] = result[0][1]
dic['task_id'] = result[0][0]
dic['comment_num'] = ''.join(re.findall('\">\s评论\[(.*?)\]\s<', soup.extract().decode(), re.S))
like = re.findall('\">赞\[(.*?)\]<', soup.extract().decode(), re.S)
if like:
dic['like_num'] = like[0]
dic['forward_num'] = ''.join(re.findall('\">转发\[(.*?)\]<', soup.extract().decode(), re.S))
create_time = soup.select('span.ct')[0].get_text().split('\xa0来自')[0]
dic['create_time'] = parsedate.parse_date(create_time)
p2 = persist_iics.Persist()
p2.insert_news(dic)
logging.info('Snatch wb news success! (Task_id: %s)' % p1_result[0][0])
p3 = persist_iics.Persist()
p3.update_task_status(p1_result[0][0])
logging.error('Snatch (Task_id: %s) failed! Updated status!' % p1_result[0][0])
def comment_prepare():
# TODO: query comment list from db.
comment_list = tuple()
p1 = persist_iics.Persist()
result = p1.query_task()
ids = re.findall('weibo.com/(.*?)/(.*?)[?]', result[0][1], re.S)[0]
url = 'http://weibo.cn/comment/{}?uid={}'.format(ids[1], ids[0])
result = dict()
result['comment'] = comment_list
result['url'] = url
return result
def comment(weibo, wb_content, wb_comment_url):
code = 1
account = weibo['usn']
password = weibo['pwd']
# service_args = [
# '--proxy=127.0.0.1:9999',
# '--proxy-type=http',
# '--ignore-ssl-errors=true'
# ]
browser = init_phantomjs_driver()
try:
browser.get("http://weibo.com")
time.sleep(3)
# browser.save_screenshot("weibocom.png")
failure = 0
while "微博-随时随地发现新鲜事" == browser.title and failure < 5:
failure += 1
username = browser.find_element_by_name("username")
pwd = browser.find_element_by_name("password")
login_submit = browser.find_element_by_class_name('W_btn_a')
username.clear()
username.send_keys(account)
pwd.clear()
pwd.send_keys(password)
login_submit.click()
time.sleep(5)
# if browser.find_element_by_class_name('verify').is_displayed():
# logger.error("Verify code is needed! (Account: %s)" % account)
if "我的首页 微博-随时随地发现新鲜事" in browser.title:
browser.get(wb_comment_url)
comment_avatar = browser.find_element_by_xpath("//div/a[@href='http://weibo.com/']")
comment_avatar.send_keys(Keys.TAB, wb_content)
time.sleep(5)
comment_submit = browser.find_element_by_xpath("//a[@class='W_btn_a']")
comment_submit.click()
time.sleep(5)
code = 0
except:
logger.error("weibo comment Unknown exception!")
raise
return code
# if __name__ == '__main__':
# print(comment({'usn': 'xxxxx', 'pwd': 'xxxxx'}, '死...死...死狗一', 'http://weibo.com/xxxxx/xxxxx'))
| [
"jraqi1994@gmail.com"
] | jraqi1994@gmail.com |
1908dbef5523b4c7baf3492790e727cfa3424bd7 | 13a1fae2f825f6c16fa15e47556e4e0b33da15b2 | /stepik course/stepik_3_1_1.py | f1d88368217a5e569d6bd9ca3bbf9a90f41a9d1a | [] | no_license | kubenet/pyLearn | 98a4a2c3243f2c03d00af82451863d65a8a38350 | 3d850de58ba525040c1f72730dcd31368766af53 | refs/heads/master | 2020-07-20T21:13:14.266289 | 2020-03-10T15:53:55 | 2020-03-10T15:53:55 | 206,710,216 | 1 | 0 | null | 2019-09-06T04:05:23 | 2019-09-06T04:05:22 | null | UTF-8 | Python | false | false | 786 | py | # Напишите функцию f(x), которая возвращает значение следующей функции, определённой на всей числовой прямой:
# 1−(x+2)^2, при x≤−2
# -x/2, при −2<x≤2
# ((x−2)^2)+1, при 2<x
#
# Требуется реализовать только функцию, решение не должно осуществлять операций ввода-вывода.
# Sample Input 1:
# 4.5
# Sample Output 1:
# 7.25
# Sample Input 2:
# -4.5
# Sample Output 2:
# -5.25
# Sample Input 3:
# 1
# Sample Output 3:
# -0.5
def f(x):
if x > 2:
return ((x-2)**2)+1
elif x <= -2:
return 1-(x+2)**2
elif (x <= 2) or (-2 < x):
return -x/2
print(f(4.5))
print(f(-4.5))
print(f(1))
| [
"kubenet@gmail.com"
] | kubenet@gmail.com |
f8bc29d25b3b055f3a4acbaac21dc04354dc7797 | e9434af6e0f542769e84a37f15860d3394367068 | /interpolator.py | 77a96c0d74c29e27f4c58e234107593baa669a9c | [] | no_license | RobinAmsters/benchmarking | 12ab77451222dfe113c0d4c66375ad5fb03d3122 | f4d7f6af804645438506a4ee5791eba596ecb52b | refs/heads/master | 2021-05-09T22:27:40.632141 | 2019-03-28T12:45:15 | 2019-03-28T12:45:15 | 118,751,992 | 0 | 2 | null | 2018-01-24T11:18:43 | 2018-01-24T10:48:03 | Python | UTF-8 | Python | false | false | 7,436 | py | import numpy as np
import sys
def transform_to_common(dataset1, dataset2):
"""
Transform both datasets to a common coordinate system, assume both datasets have the same amount of points.
:param dataset1:
:param dataset2:
:return:
"""
keep = np.invert(np.bitwise_or(np.isnan(np.sum(dataset1, axis=1)), np.isnan(np.sum(dataset2, axis=1))))
dataset1 = dataset1[keep]
dataset2 = dataset2[keep]
mean1 = np.mean(dataset1, axis=0)
mean2 = np.mean(dataset2, axis=0)
zeroAvg1 = dataset1 - mean1
zeroAvg2 = dataset2 - mean2
R = extract_rotation_matrix(zeroAvg1, zeroAvg2)
tranformed1 = np.dot(R, zeroAvg1.transpose()).transpose()
translation = np.mean(dataset2 - np.dot(R, dataset1.transpose()).transpose(), axis=0)
return tranformed1, zeroAvg2, R, translation
def extract_rotation_matrix(zeroAveragedPositions1, zeroAveragedPositions2):
"""
Find the rotation matrix between two point clouds with a mean position of 0. The rotatation matrix is defined as follows:
R . zeroAveragedPositions1 == zeroAveragedPositions2 + epsilon
:param zeroAveragedPositions1: The first point cloud
:param zeroAveragedPositions2: The second point cloud
:return: The rotation matrix
"""
u, s, v_trans = np.linalg.svd(np.dot(zeroAveragedPositions1.transpose(), zeroAveragedPositions2))
d = np.linalg.det(np.dot(v_trans.transpose(), u.transpose()))
R = np.dot(np.dot(v_trans.transpose(), np.array([[1, 0, 0], [0, 1, 0], [0, 0, d]])), u.transpose())
return R
def find_pose(originalPositions, movedPositions):
"""
Find the pose matrix for a given set of marker positions on the robot and the coordinates of these points for which to calculate the pose matrix.
:param originalPositions: The positions of the markers on the robot (rows are different markers).
:param movedPositions: The moved positions (rows are different markers).
:return: The pose matrix.
"""
if len(originalPositions) != len(movedPositions):
raise Exception(
"To find the pose of the robot, the same amount of markers have to be defined as the amount of measured markers")
indices = np.invert(np.isnan(np.sum(movedPositions, axis=1)))
if sum(indices) < 3:
return None
originalPositions = np.array(originalPositions)[indices, :]
movedPositions = movedPositions[indices, :]
zeroAvg1 = originalPositions - np.mean(originalPositions, axis=0)
zeroAvg2 = movedPositions - np.mean(movedPositions, axis=0)
R = extract_rotation_matrix(zeroAvg1, zeroAvg2)
transformed_input = np.dot(R, originalPositions.transpose()).transpose()
meanVec = np.mean(transformed_input, axis=0)
meanPos = np.mean(movedPositions, axis=0) - meanVec
meanPos.shape = (3, 1)
pose = np.block([[R, meanPos], [0, 0, 0, 1]])
error = 0
for i in range(originalPositions.shape[0]):
error += np.sum(np.abs(movedPositions[i, :] - np.dot(pose, np.append(originalPositions[i, :], 1))[:3]))
if error > 30:
return None
return np.block([[R, meanPos], [0, 0, 0, 1]]) # TODO: Not sure about block
def overlap_datasets(markerTimePoints, transformedMarkerPositions, evaluationTrack, referenceFramerate=1/50):
"""
:param initialMarkerPositions:
:param markerTracks: The measured positions of the different markers, the measurements are assumed to be at a constant rate.
The timestamps are assumed to be equal among the different markers.
:param initialEvaluationPosition:
:param evaluationTrack:
:return:
"""
evaluationTrack[:, 0] -= evaluationTrack[0, 0]
kryptonIndices = get_virtual_indices(referenceFramerate, evaluationTrack[:, 0])
return timeShift(evaluationTrack, kryptonIndices, transformedMarkerPositions, markerTimePoints)
def get_transformed_marker_position(initialEvaluationPosition, initialMarkerPositions, markerTracks):
transformedMarkerPositions = list()
for i in range(markerTracks.shape[1]):
pose = find_pose(initialMarkerPositions, markerTracks[:, i, 1:])
if pose is not None:
transformedMarkerPositions.append(np.dot(pose, np.append(initialEvaluationPosition, 1))[:3])
else:
transformedMarkerPositions.append(np.array([np.nan, np.nan, np.nan]))
transformedMarkerPositions = np.array(transformedMarkerPositions)
return transformedMarkerPositions
def get_transformed_vive_position(initialEvaluationPosition, initialVivePosition, viveMatrix):
transformedVivePositions = list()
for i in range(len(viveMatrix)):
R = viveMatrix[i][0][:, :3]
p = viveMatrix[i][0][:, 3] * 1000
originP = p - np.dot(R, initialVivePosition) + initialVivePosition
transformedVivePositions.append(originP + np.dot(R, initialEvaluationPosition))
return transformedVivePositions
def timeShift(evaluationTrack, kryptonIndices, transformedMarkerPositions, markerTimePoints):
"""
Find how much the evaluation track is shifted in time versus the reference (transformedMarkerPOsitions)
:param evaluationTrack: numpy ndarray containing the positions of the system to evaluate: [[time, X, Y, Z]]
:param kryptonIndices: a list of numbers containing the matching index in transformedMarkerPositions for each point in the evaluationtrack.
:param transformedMarkerPositions: the positions of the measurement system to evaluate according to the Krypton system.
:param markerTimePoints: the timestamps of transformedMarkerPositions
:return:
"""
minOverlap = 5 # seconds of overlap TODO: figure out realistic overlap
minErr = sys.float_info.max
finalOffset = 0
finalR = None
finalTranslation = None
minOffset = - kryptonIndices[np.searchsorted(evaluationTrack[:, 0], evaluationTrack[-1, 0] - minOverlap)]
maxOffset = np.searchsorted(markerTimePoints, markerTimePoints[-1] - minOverlap)
offsets = range(minOffset, maxOffset)
for offset in offsets:
start = np.searchsorted(kryptonIndices + offset, 0, side='right')
end = np.searchsorted(kryptonIndices + offset, len(transformedMarkerPositions))
markerPoints = transformedMarkerPositions[kryptonIndices[start:end] + offset]
evaluationPoints = evaluationTrack[start:end]
evaluationTransformed, markerTransformed, R, translation = transform_to_common(evaluationPoints[:, 1:],
markerPoints)
error = error_func(markerTransformed, evaluationTransformed)
if error < minErr:
minErr = error
finalOffset = offset
finalR = R
finalTranslation = translation
return finalOffset, finalR, finalTranslation
def get_virtual_indices(kryptonRate, timestamps):
return np.array(map(lambda a: int(round(a * kryptonRate)), timestamps))
def error_func(referenceSet, evaluationSet):
refMean = np.mean(referenceSet, axis=0)
variance = np.sum((referenceSet - refMean) ** 2)
if variance ** 0.5 < 2000: # guarantee at least 5m of movement within interval.
return sys.float_info.max
return float(np.sum(np.abs(referenceSet - evaluationSet))) / len(referenceSet)
if __name__ == '__main__':
c = np.array([[1, 2, 5], [3, 5, 4], [10, 8, 9]])
m = np.array([[2, -1, 5], [5, -3, 4], [8, -10, 9]])
pose = find_pose(c, m)
print(pose)
| [
"quinten.lauwers1@student.kuleuven.be"
] | quinten.lauwers1@student.kuleuven.be |
aea9bbdf52c15b1029619192896682d0da4103c2 | a4b46342bc37d2d08b19934c5230928575d9cc39 | /getEmailByRegex.py | 7ad1f201c2b92c1916a084f220d32e11906a7fd5 | [] | no_license | yenchenhuang/regex_practice | 5daec1e13903ad233b2d67dba145593e4a0e7966 | 59f138639f783198d813572f2487566955355d0b | refs/heads/master | 2020-03-26T05:34:41.882199 | 2016-06-22T09:13:52 | 2016-06-22T09:13:52 | 144,563,581 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 396 | py | import re
def is_email(input):
pattern = r"([\w._]+)@([\w_\-.]+?)"
match = re.match(pattern, input)
if match:
return True
else:
return False
def get_emails(paragraph):
pattern = r"[\w._]+@[\w_\-.]+"
matches = re.findall(pattern, paragraph)
return matches
def get_accounts(paragraph):
pattern = r"([\w._]+)@[\w_\-.]+"
matches = re.findall(pattern, paragraph)
return matches
| [
"yenchenhuang@kkbox.com"
] | yenchenhuang@kkbox.com |
b477a9cee5d1d50b4effb6e86d254fa10629c6f1 | d63811f9944dead8a745a46e1382f64800c72c5e | /linuxYazKampı/sonuç/petimebak/adverts/views.py | ea08662d7d76a0d3d26ba3224ffcc6795ba666ad | [] | no_license | Arciles/Notes | 9dd77425209b9a10a6503dcd27a5c48c9666c35b | 095e361bdb11ca72c3bff801ed4a9b938827c84a | refs/heads/master | 2020-12-11T05:46:11.416975 | 2014-11-22T12:08:37 | 2014-11-22T12:08:37 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,924 | py | from datetime import datetime
from django.contrib.auth.decorators import login_required
from django.core.urlresolvers import reverse
from django.shortcuts import render_to_response, get_object_or_404, redirect
from django.template import RequestContext
from adverts.forms import AdvertCreationForm, PhotoCreationForm
from adverts.models import Advert
from messages.forms import NewMessageForm
def home(request):
adverts = (Advert.objects
.filter(is_published=True, end_date__gte=datetime.now())
.order_by("-date_created"))
return render_to_response("index.html", {
"adverts": adverts
}, RequestContext(request))
@login_required
def new_advert(request):
form = AdvertCreationForm()
success = False
if request.method == "POST":
form = AdvertCreationForm(request.POST)
if form.is_valid():
form.instance.user = request.user
form.save()
success = True
return render_to_response("new_advert.html", {
"form": form,
"success": success,
}, RequestContext(request))
def detail_advert(request, pk):
advert = get_object_or_404(Advert, id=pk)
message_sent = request.GET.get("message_sent")
form = NewMessageForm()
return render_to_response("detail.html", {
"advert": advert,
"form": form,
"message_sent": message_sent
}, RequestContext(request))
def photo_add(request, pk):
advert = get_object_or_404(Advert, id=pk)
form = PhotoCreationForm()
if request.method == "POST":
form = PhotoCreationForm(request.POST, request.FILES)
if form.is_valid():
form.instance.advert = advert
form.save()
return redirect(reverse('detail_advert', args=[pk]))
return render_to_response("photo_add.html", {
"advert": advert,
"form": form,
}, RequestContext(request))
| [
"salihklc91@gmail.com"
] | salihklc91@gmail.com |
b6d4b00e9ba7fd2e1ffb15551e74584d5f265b5d | ed61c386fbe2ab18a73e6c4ac4c581540638dba6 | /src/old verion of code/propagation.py | d0295ff5fd4ade51e1ce136c9a8ddc81f37f1650 | [] | no_license | allenqaq/Online-Social-Network | 5f244cbf3b9f53fe68a926847e485c3a5ab9782e | c6f863827521ed787cf0120c3e02ae0485004a6a | refs/heads/master | 2020-03-08T17:34:54.384051 | 2018-12-21T08:42:34 | 2018-12-21T08:42:34 | 128,272,678 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,024 | py | '''
Created on Mar 5, 2017
@author: allen
'''
from math import sqrt
import numpy as np
# def mean(lst):
# """calculates mean"""
# sum = 0
# for i in range(len(lst)):
# sum += lst[i]
# return (sum / len(lst))
# def stddev(lst):
# """calculates standard deviation"""
# sum = 0
# mn = mean(lst)
# for i in range(len(lst)) :
# sum += pow((lst[i]-mn),2)
# return sqrt(sum/len(lst)-1)
def eccenricity(items) :
if not items :
return 0;
sd = np.std(items)
# print(sd)
if sd == 0 :
return 0
max1 = max(items)
items.remove(max1)
if len(items) != 0 :
max2 = max(items)
else :
max2 = 0
# print((max1 - max2) / sd)
return (max1 - max2) / sd
def matchScores(lgraph, rgraph, mapping, lnode) :
# print('match scores start')
scores = {}
# for i in lgraph.nodes():
# scores[i] = 0
# print(scores)
# scores[lnode] = 0.1
listIn1 = lgraph.in_edges(lnode)
for lnbr in listIn1 :
#lnbr is like (1111, 1112)
if lnbr[0] not in mapping[0].keys() :
continue
rnbr = mapping[0][lnbr[0]]
listOut1 = rgraph.out_edges(rnbr)
for rnode in listOut1 :
#rnode is like (1111, 1112)
# if rnode[1] in mapping[0].keys() or rgraph.in_degree(rnode[1]) > 100 or rgraph.in_degree(rnode[1]) == 1:
if rnode[1] in mapping[0].keys() :
continue
else :
skip = rgraph.in_degree(rnode[1]) - rgraph.in_degree(lnode)
# skip is a degree check supposing 2 mapping mode degree differential is less than 3
if skip > 3 or skip < -3 :
continue
elif rnode[1] in scores.keys() :
scores[rnode[1]] += 1 / sqrt(rgraph.in_degree(rnode[1]))
else :
scores[rnode[1]] = 1 / sqrt(rgraph.in_degree(rnode[1]))
listOut2 = lgraph.out_edges(lnode)
for lnbr in listOut2 :
#lnbr is like (1111, 1112)
if lnbr[1] not in mapping[0].keys() :
continue
rnbr = mapping[0][lnbr[1]]
listIn2 = rgraph.in_edges(rnbr)
for rnode in listIn2 :
#rnode is like (1111, 1112)
# if rnode[0] in mapping[0].keys() or rgraph.out_degree(rnode[0]) > 100 or rgraph.out_degree(rnode[0]) == 1:
if rnode[0] in mapping[0].keys() :
continue
else :
skip = rgraph.out_degree(rnode[0]) - rgraph.out_degree(lnode)
# skip is a degree check supposing 2 mapping mode degree differential is less than 3
if skip > 3 or skip < -3 :
continue
if rnode[0] in scores.keys() :
scores[rnode[0]] += 1 / sqrt(rgraph.out_degree(rnode[0]))
else :
scores[rnode[0]] = 1 / sqrt(rgraph.out_degree(rnode[0]))
# if lnode in scores.keys() and scores[lnode] != 0 :
# print("lnode :"),
# print(lnode)
# print("shoule be scores : ")
# print(scores[lnode])
# print("max id is :"),
# print(max(scores.items(), key=lambda x: x[1])[0])
# print(max(scores.values()))
# print(scores)
# # for (k,v) in scores.items() :
# # if v == max(scores.values()) :
# # print(k),
# # print
# print("--------------------------------------")
return scores
def propagationStep(lgraph, rgraph, mapping) :
scores = {}
node_acount = 0
for lnode in lgraph.nodes() :
node_acount = node_acount + 1
data_len = len(lgraph.nodes())
rate = node_acount * 100.0 / data_len
print('-------'),
print(rate),
print(' %')
if lnode in mapping[0].keys() :
continue
scores[lnode] = matchScores(lgraph, rgraph, mapping, lnode)
# print(scores[lnode])
if eccenricity(scores[lnode].values()) < 0.01 :
# 0.01 is theta, a parameter that controls the tradeoff between the yield and the accuracy.
continue
rnode = max(scores[lnode].items(), key=lambda x: x[1])[0]
scores[rnode] = matchScores(rgraph, lgraph, mapping, rnode)
# no need to invert mapping
if eccenricity(scores[rnode].values()) < 0.01 :
# 0.01 is theta, a parameter that controls the tradeoff between the yield and the accuracy.
continue
reverse_match = max(scores[rnode].items(), key=lambda x: x[1])[0]
print("reverse_match :"),
print(reverse_match)
print("lnode :"),
print(lnode)
# print(scores[rnode][reverse_match])
# print(scores[rnode][lnode])
print("======================================")
if reverse_match != lnode :
continue
else :
mapping[0][lnode] = rnode
| [
"allenqaq555@gmail.com"
] | allenqaq555@gmail.com |
69163c15593175ec5108d8614e170be4b086b0cc | d5a84ba1417d59d6b8eff26124a37ba7186d7e33 | /test_calculator.py | a303fcf1954dd8cadf456028cef326d495242bd5 | [] | no_license | jonathanzerox/tdd-python | 7c4ce49acb6eec562c9a382e9aab83cd1963aa23 | cd718ecdd646046dd0e2d437d050d75f16378280 | refs/heads/master | 2021-08-23T17:39:47.264755 | 2017-12-05T22:54:25 | 2017-12-05T22:54:25 | 113,240,878 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 476 | py | import unittest
from calculator import Calculator
class TestCalculator(unittest.TestCase):
def SetUp():
print("Setting things up")
def TearDown():
print("Releasing allocated resources back")
def test_addition(self):
calc = Calculator()
self.assertEqual(4, calc.add(2, 2))
def test_multiplication(self):
calc = Calculator()
self.assertEqual(8, calc.mul(4, 2))
if __name__ == '__main__':
unittest.main()
| [
"jonathanzerox@hotmail.com"
] | jonathanzerox@hotmail.com |
7ac445a7981cc09e31bfafce07f08ab38310efce | 2a1146bd74be4ae270bd2dc105e1917aa13a3bfe | /gotti/modules/modules.py | 346c86de735720927666368efd616105af0b2a46 | [
"MIT"
] | permissive | HellBringerReal/Telegram-Bot | 0a2721ed04667c6bb1347f7ccdf0e657a360a71a | c204de5e8212fd32aaae6afd92c2bc7999457d4f | refs/heads/master | 2023-06-28T20:34:50.909101 | 2021-08-11T20:22:27 | 2021-08-11T20:22:27 | 290,599,195 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,947 | py | import importlib
from telegram import Bot, Update, ParseMode
from telegram.ext import CommandHandler, run_async
from gotti import dispatcher
from gotti.__main__ import (IMPORTED, HELPABLE, MIGRATEABLE, STATS, USER_INFO, DATA_IMPORT, DATA_EXPORT, CHAT_SETTINGS,
USER_SETTINGS)
from gotti.modules.helper_funcs.chat_status import sudo_plus, dev_plus
@run_async
@dev_plus
def load(bot: Bot, update: Update):
message = update.effective_message
text = message.text.split(" ", 1)[1]
load_messasge = message.reply_text(f"Attempting to load module : <b>{text}</b>", parse_mode=ParseMode.HTML)
try:
imported_module = importlib.import_module("gotti.modules." + text)
except:
load_messasge.edit_text("Does that module even exist?")
return
if not hasattr(imported_module, "__mod_name__"):
imported_module.__mod_name__ = imported_module.__name__
if not imported_module.__mod_name__.lower() in IMPORTED:
IMPORTED[imported_module.__mod_name__.lower()] = imported_module
else:
load_messasge.edit_text("Module already loaded.")
return
if "__handlers__" in dir(imported_module):
handlers = imported_module.__handlers__
for handler in handlers:
if type(handler) != tuple:
dispatcher.add_handler(handler)
else:
handler_name, priority = handler
dispatcher.add_handler(handler_name, priority)
else:
IMPORTED.pop(imported_module.__mod_name__.lower())
load_messasge.edit_text("The module cannot be loaded.")
return
if hasattr(imported_module, "__help__") and imported_module.__help__:
HELPABLE[imported_module.__mod_name__.lower()] = imported_module
# Chats to migrate on chat_migrated events
if hasattr(imported_module, "__migrate__"):
MIGRATEABLE.append(imported_module)
if hasattr(imported_module, "__stats__"):
STATS.append(imported_module)
if hasattr(imported_module, "__user_info__"):
USER_INFO.append(imported_module)
if hasattr(imported_module, "__import_data__"):
DATA_IMPORT.append(imported_module)
if hasattr(imported_module, "__export_data__"):
DATA_EXPORT.append(imported_module)
if hasattr(imported_module, "__chat_settings__"):
CHAT_SETTINGS[imported_module.__mod_name__.lower()] = imported_module
if hasattr(imported_module, "__user_settings__"):
USER_SETTINGS[imported_module.__mod_name__.lower()] = imported_module
load_messasge.edit_text("Successfully loaded module : <b>{}</b>".format(text), parse_mode=ParseMode.HTML)
@run_async
@dev_plus
def unload(bot: Bot, update: Update):
message = update.effective_message
text = message.text.split(" ", 1)[1]
unload_messasge = message.reply_text(f"Attempting to unload module : <b>{text}</b>", parse_mode=ParseMode.HTML)
try:
imported_module = importlib.import_module("gotti.modules." + text)
except:
unload_messasge.edit_text("Does that module even exist?")
return
if not hasattr(imported_module, "__mod_name__"):
imported_module.__mod_name__ = imported_module.__name__
if imported_module.__mod_name__.lower() in IMPORTED:
IMPORTED.pop(imported_module.__mod_name__.lower())
else:
unload_messasge.edit_text("Can't unload something that isn't loaded.")
return
if "__handlers__" in dir(imported_module):
handlers = imported_module.__handlers__
for handler in handlers:
if type(handler) == bool:
unload_messasge.edit_text("This module can't be unloaded!")
return
elif type(handler) != tuple:
dispatcher.remove_handler(handler)
else:
handler_name, priority = handler
dispatcher.remove_handler(handler_name, priority)
else:
unload_messasge.edit_text("The module cannot be unloaded.")
return
if hasattr(imported_module, "__help__") and imported_module.__help__:
HELPABLE.pop(imported_module.__mod_name__.lower())
# Chats to migrate on chat_migrated events
if hasattr(imported_module, "__migrate__"):
MIGRATEABLE.remove(imported_module)
if hasattr(imported_module, "__stats__"):
STATS.remove(imported_module)
if hasattr(imported_module, "__user_info__"):
USER_INFO.remove(imported_module)
if hasattr(imported_module, "__import_data__"):
DATA_IMPORT.remove(imported_module)
if hasattr(imported_module, "__export_data__"):
DATA_EXPORT.remove(imported_module)
if hasattr(imported_module, "__chat_settings__"):
CHAT_SETTINGS.pop(imported_module.__mod_name__.lower())
if hasattr(imported_module, "__user_settings__"):
USER_SETTINGS.pop(imported_module.__mod_name__.lower())
unload_messasge.edit_text(f"Successfully unloaded module : <b>{text}</b>", parse_mode=ParseMode.HTML)
@run_async
@sudo_plus
def listmodules(bot: Bot, update: Update):
message = update.effective_message
module_list = []
for helpable_module in HELPABLE:
helpable_module_info = IMPORTED[helpable_module]
file_info = IMPORTED[helpable_module_info.__mod_name__.lower()]
file_name = file_info.__name__.rsplit("gotti.modules.", 1)[1]
mod_name = file_info.__mod_name__
module_list.append(f'- <code>{mod_name} ({file_name})</code>\n')
module_list = "Following modules are loaded : \n\n" + ''.join(module_list)
message.reply_text(module_list, parse_mode=ParseMode.HTML)
LOAD_HANDLER = CommandHandler("load", load)
UNLOAD_HANDLER = CommandHandler("unload", unload)
LISTMODULES_HANDLER = CommandHandler("listmodules", listmodules)
dispatcher.add_handler(LOAD_HANDLER)
dispatcher.add_handler(UNLOAD_HANDLER)
dispatcher.add_handler(LISTMODULES_HANDLER)
__mod_name__ = "MODULES"
| [
"noreply@github.com"
] | HellBringerReal.noreply@github.com |
a852d4447c5f5e6261198b26f9281cce50269c1c | d393e865b83edc1b83fe80b716775b8036c51af4 | /Preprocessing.py | 40094cabfa374078ce04b111e05fd9cab6011fa5 | [] | no_license | dwaydwaydway/KKStream-Deep-Learning-Workshop | 1137724577c46b9d0a039a8b64f84abf0a0ea91f | 7c9da5114b6901052d479228fa40a5628e646e25 | refs/heads/master | 2020-05-25T04:44:29.691746 | 2020-03-11T09:25:25 | 2020-03-11T09:25:25 | 187,633,348 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,497 | py | import numpy as np
import pandas as pd
import multiprocessing as mp
from datetime import datetime
import tqdm
import pickle
import csv
import math
import warnings
import json
warnings.filterwarnings("ignore")
def main(config):
pool = mp.Pool(processes=config["n_workers"])
pool.map(job, range(1, 76))
# Return one hot encoding of length = base and the nth element = 1
def make_onehot(n, base):
if n == -1 :
return list(np.zeros(base))
onehot = np.zeros(base)
onehot[n-1] = 1
return list(onehot)
# Return the time slot number of the given datatime boject
def is_time_slot(datetime, time_slot_comp):
slot = 0
for i in range(4):
if datetime.time() >= time_slot_comp[i][0] and datetime.time() < time_slot_comp[i][1]:
slot = i
return slot
# Return [week, time slot, time slot of the week]
def time_slot(datetime, begin_datetime, end_datetime, time_slot_comp):
if datetime < begin_datetime or datetime > end_datetime:
return [-1, -1, -1]
diff = datetime - begin_datetime
time_slot = is_time_slot(datetime, time_slot_comp)
return [int(diff.days / 7), diff.days*4 + time_slot, datetime.weekday()*4 + time_slot]
# Return the scaleed played duration value
def scale_played_duration(n):
return math.log(1 + n / 9550675.0)
# Return the Processed data
def prepare_data(row, platform):
temp = list()
# platform
temp += make_onehot(platform[row[4]], 3)
# connection type
if 'wifi' in row[7]:
connect = make_onehot(0, 3)
elif 'cellular' in row[7]:
connect = make_onehot(1, 3)
elif 'online' in row[7]:
connect = make_onehot(2, 3)
else:
connect = make_onehot(-1, 3)
temp += connect
# watch ratio
temp.append(float(row[5]) / (float(row[6]) + 1e-10))
# total number of episode
temp.append(math.log(1 + float(row[5]) / 210.0))
# limit playzone countdown
temp.append(1 if 'limit playzone countdown' in row[3] else 0)
# error
temp.append(1 if 'error' in row[3] else 0)
# (video ended, program stopped or enlarged-reduced, program stopped)
temp.append(1 if 'ed' in row[3] else 0)
# played duration of popular title
temp.append(math.log(1 + float(row[2]) / 5224.0) if int(row[0]) in [74, 79, 77] else 0)
return temp
# Extact selected features from data files(This part could be a bit messy)
def job(k):
with open("preprocessing_config.json") as f:
config = json.load(f)
begin_datetime = datetime.strptime('2017-01-02 01:00:00.00', '%Y-%m-%d %H:%M:%S.%f')
end_datetime = datetime.strptime('2017-08-14 01:00:00.00', '%Y-%m-%d %H:%M:%S.%f')
time_slot_comp = [[datetime.strptime('01:00:00', '%H:%M:%S').time(), datetime.strptime('09:00:00', '%H:%M:%S').time()],
[datetime.strptime('09:00:0', '%H:%M:%S').time(), datetime.strptime('17:00:00', '%H:%M:%S').time()],
[datetime.strptime('17:00:0', '%H:%M:%S').time(), datetime.strptime('21:00:00', '%H:%M:%S').time()],
[datetime.strptime('21:00:00', '%H:%M:%S').time(), datetime.strptime('01:00:00', '%H:%M:%S').time()]]
platform = { 'Web': 0, 'iOS': 1, 'Android': 2 }
file = pd.read_csv((config["data_folder"] + "/data-0{:0>2d}.csv").format(k))
n = file.groupby(['user_id']).size().values
file = file.drop(columns=['user_id', 'device_id', 'session_id', 'is_trailer'])
file = np.split(np.asarray(file.values), np.add.accumulate(n), axis=0)[:len(n)]
pad = config["pad_token"]
collect = []
for idx in file:
person_data = np.ones((32, 28, 13)) * pad
week_data = np.ones((28, 13)) * pad
time_slot_data = []
watch_time_sum = 0
prev = time_slot(datetime.strptime(idx[0][1], '%Y-%m-%d %H:%M:%S.%f'), begin_datetime, end_datetime, time_slot_comp)
for row in idx:
now = time_slot(datetime.strptime(row[1], '%Y-%m-%d %H:%M:%S.%f'), begin_datetime, end_datetime, time_slot_comp)
if prev[0] == -1:
prev = now
if now[0] == -1:
continue
if now[1] == prev[1]:
time_slot_data.append(prepare_data(row, platform))
watch_time_sum += float(row[2])
else:
time_slot_data = np.mean(time_slot_data, axis=0).tolist()
time_slot_data.append(scale_played_duration(watch_time_sum))
week_data[prev[2]] = time_slot_data
time_slot_data = []
if prev[0] < now[0]:
person_data[prev[0]] = week_data
week_data = week_data * 0 + pad
time_slot_data.append(prepare_data(row, platform))
watch_time_sum += float(row[2])
prev = now
if len(time_slot_data) != 0:
time_slot_data = np.mean(time_slot_data, axis=0).tolist()
time_slot_data.append(scale_played_duration(watch_time_sum))
week_data[prev[2]] = time_slot_data
person_data[prev[0]] = week_data
collect.append(person_data)
with open((config["preprocessed_data_folder"] +"/data-00{:0>2d}_preprocessed.pkl").format(k), 'wb') as handle:
pickle.dump(collect, handle)
if __name__ == '__main__':
with open("preprocessing_config.json") as f:
config = json.load(f)
main(config) | [
"tim8733123@gmail.com"
] | tim8733123@gmail.com |
c6c061e8c03a18c82ab3cd76a3c826f444b900f1 | 14c831041866edeed1fb3a5643405542fe736939 | /datasetGen.py | f1e20ec710d378afc9b9e20cff19a3c148fd8dba | [] | no_license | Rumee040/Retinal-Vessel-Segmentation-using-Convolutional-Network | d4a37d72d8fde637b53c5d2e279c9848d1dc1313 | 41333730fd2c27a99c0bbb149dfaa739c04ece50 | refs/heads/master | 2021-05-06T06:01:16.863817 | 2017-12-20T11:58:19 | 2017-12-20T11:58:19 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,468 | py | #import cv2
import numpy as np
from sys import exit
from PIL import Image
import matplotlib.pyplot as plt
import random
## defining parameters
patchSize = 27 # rectangular patch with size patchSize*patchSize*channel
patchPerImg = 1000 # patches per image
numImage = 20 # number of images
totalPatch = patchPerImg * numImage
data = np.ones((totalPatch, patchSize, patchSize, 3)) # all of the patches will be stored here
dataLoc = np.ones((totalPatch, 2)) # location of the patches stores as (row, column)
dataLabel = np.ones((totalPatch)) # label of the patches 0 - neg, 1 - pos
balance = 0.5 # balance between positive and negative patches
positive = int(patchPerImg * 0.5) # number of positive image in an image
negative = patchPerImg - positive # number of negative image in an image
## reading the imageand mask
for i in range(1, numImage + 1):
imgNum = str(i)
if i < 10:
imgNum = '0' + imgNum
imgName = imgNum + '_test.tif'
img = Image.open('E:\\library of EEE\\4-2\\eee 426\\data\\DRIVE\\DRIVE\\test\\images\\' + imgName)
maskName = imgNum + '_test_mask.gif'
mask = Image.open('E:\\library of EEE\\4-2\\eee 426\\data\\DRIVE\\DRIVE\\test\\mask\\' + maskName)
gndTruthName = imgNum + '_manual1.gif'
gndTruth = Image.open('E:\\library of EEE\\4-2\\eee 426\\data\\DRIVE\\DRIVE\\test\\1st_manual\\' + gndTruthName)
## converting them to numpy array
img = np.array(img)
#img = np.array(img.getdata()).reshape(img.size[1], img.size[0], 3) # Image class store image as (width, height) but we want it as (row, column)
#img = img.astype('float32') / 255 # to see the image in plt
mask = mask.convert('RGB')
#mask = np.array(mask.getdata()).reshape(mask.size[1], mask.size[0], 3)
mask = np.array(mask)
#mask = mask.astype('float32') / 255
gndTruth = gndTruth.convert('RGB')
gndTruth = np.array(gndTruth)[:,:,0]
#gndTruth = gndTruth.astype('float32') / 255
## cutting out patches from the image
imgRow = img.shape[0]
imgCol = img.shape[1]
count = 0
ind = (i - 1) * patchPerImg
posCount = 0
negCount = 0
while count < patchPerImg:
r = int(round(random.uniform(0, img.shape[0])))
c = int(round(random.uniform(0, img.shape[1])))
rStart = r - patchSize // 2
rEnd = r + patchSize // 2 + 1
cStart = c - patchSize // 2
cEnd = c + patchSize // 2 + 1
if np.all(mask[rStart:rEnd, cStart:cEnd]) and r > 13 and r < imgRow - 14 and c > 13 and c < imgCol - 14:
label = gndTruth[r, c]
if label == 0:
if negCount == negative:
continue
else:
negCount += 1
else:
if posCount == positive:
continue
else:
posCount += 1
data[ind + count, :, :, :] = img[rStart:rEnd, cStart:cEnd, :]
dataLoc[ind + count] = np.array([r, c])
dataLabel[ind + count] = label
count += 1
print(negCount, posCount)
print(np.count_nonzero(dataLabel))
## storing the images and data
np.save('E:\\library of EEE\\4-2\\eee 426\\data\\MSCprojectDataBase\\simpleClassifierDataBase\\DRIVEtestData', data)
np.save('E:\\library of EEE\\4-2\\eee 426\\data\\MSCprojectDataBase\\simpleClassifierDataBase\\DRIVEtestDataLcation', dataLoc)
np.save('E:\\library of EEE\\4-2\\eee 426\\data\\MSCprojectDataBase\\simpleClassifierDataBase\\DRIVEtestDataLabel', dataLabel) | [
"noreply@github.com"
] | Rumee040.noreply@github.com |
2a1f5e4881e26548e1ce7fdb9043a6c590f91749 | 90419da201cd4948a27d3612f0b482c68026c96f | /sdk/python/pulumi_azure_nextgen/devtestlab/latest/get_lab.py | f41e214f30cdd3cfa752fbedcad88a1a4eccd182 | [
"BSD-3-Clause",
"Apache-2.0"
] | permissive | test-wiz-sec/pulumi-azure-nextgen | cd4bee5d70cb0d332c04f16bb54e17d016d2adaf | 20a695af0d020b34b0f1c336e1b69702755174cc | refs/heads/master | 2023-06-08T02:35:52.639773 | 2020-11-06T22:39:06 | 2020-11-06T22:39:06 | 312,993,761 | 0 | 0 | Apache-2.0 | 2023-06-02T06:47:28 | 2020-11-15T09:04:00 | null | UTF-8 | Python | false | false | 16,776 | py | # coding=utf-8
# *** WARNING: this file was generated by the Pulumi SDK Generator. ***
# *** Do not edit by hand unless you're certain you know what you are doing! ***
import warnings
import pulumi
import pulumi.runtime
from typing import Any, Mapping, Optional, Sequence, Union
from ... import _utilities, _tables
from . import outputs
__all__ = [
'GetLabResult',
'AwaitableGetLabResult',
'get_lab',
]
@pulumi.output_type
class GetLabResult:
"""
A lab.
"""
def __init__(__self__, announcement=None, artifacts_storage_account=None, created_date=None, default_premium_storage_account=None, default_storage_account=None, environment_permission=None, extended_properties=None, lab_storage_type=None, load_balancer_id=None, location=None, mandatory_artifacts_resource_ids_linux=None, mandatory_artifacts_resource_ids_windows=None, name=None, network_security_group_id=None, premium_data_disk_storage_account=None, premium_data_disks=None, provisioning_state=None, public_ip_id=None, support=None, tags=None, type=None, unique_identifier=None, vault_name=None, vm_creation_resource_group=None):
if announcement and not isinstance(announcement, dict):
raise TypeError("Expected argument 'announcement' to be a dict")
pulumi.set(__self__, "announcement", announcement)
if artifacts_storage_account and not isinstance(artifacts_storage_account, str):
raise TypeError("Expected argument 'artifacts_storage_account' to be a str")
pulumi.set(__self__, "artifacts_storage_account", artifacts_storage_account)
if created_date and not isinstance(created_date, str):
raise TypeError("Expected argument 'created_date' to be a str")
pulumi.set(__self__, "created_date", created_date)
if default_premium_storage_account and not isinstance(default_premium_storage_account, str):
raise TypeError("Expected argument 'default_premium_storage_account' to be a str")
pulumi.set(__self__, "default_premium_storage_account", default_premium_storage_account)
if default_storage_account and not isinstance(default_storage_account, str):
raise TypeError("Expected argument 'default_storage_account' to be a str")
pulumi.set(__self__, "default_storage_account", default_storage_account)
if environment_permission and not isinstance(environment_permission, str):
raise TypeError("Expected argument 'environment_permission' to be a str")
pulumi.set(__self__, "environment_permission", environment_permission)
if extended_properties and not isinstance(extended_properties, dict):
raise TypeError("Expected argument 'extended_properties' to be a dict")
pulumi.set(__self__, "extended_properties", extended_properties)
if lab_storage_type and not isinstance(lab_storage_type, str):
raise TypeError("Expected argument 'lab_storage_type' to be a str")
pulumi.set(__self__, "lab_storage_type", lab_storage_type)
if load_balancer_id and not isinstance(load_balancer_id, str):
raise TypeError("Expected argument 'load_balancer_id' to be a str")
pulumi.set(__self__, "load_balancer_id", load_balancer_id)
if location and not isinstance(location, str):
raise TypeError("Expected argument 'location' to be a str")
pulumi.set(__self__, "location", location)
if mandatory_artifacts_resource_ids_linux and not isinstance(mandatory_artifacts_resource_ids_linux, list):
raise TypeError("Expected argument 'mandatory_artifacts_resource_ids_linux' to be a list")
pulumi.set(__self__, "mandatory_artifacts_resource_ids_linux", mandatory_artifacts_resource_ids_linux)
if mandatory_artifacts_resource_ids_windows and not isinstance(mandatory_artifacts_resource_ids_windows, list):
raise TypeError("Expected argument 'mandatory_artifacts_resource_ids_windows' to be a list")
pulumi.set(__self__, "mandatory_artifacts_resource_ids_windows", mandatory_artifacts_resource_ids_windows)
if name and not isinstance(name, str):
raise TypeError("Expected argument 'name' to be a str")
pulumi.set(__self__, "name", name)
if network_security_group_id and not isinstance(network_security_group_id, str):
raise TypeError("Expected argument 'network_security_group_id' to be a str")
pulumi.set(__self__, "network_security_group_id", network_security_group_id)
if premium_data_disk_storage_account and not isinstance(premium_data_disk_storage_account, str):
raise TypeError("Expected argument 'premium_data_disk_storage_account' to be a str")
pulumi.set(__self__, "premium_data_disk_storage_account", premium_data_disk_storage_account)
if premium_data_disks and not isinstance(premium_data_disks, str):
raise TypeError("Expected argument 'premium_data_disks' to be a str")
pulumi.set(__self__, "premium_data_disks", premium_data_disks)
if provisioning_state and not isinstance(provisioning_state, str):
raise TypeError("Expected argument 'provisioning_state' to be a str")
pulumi.set(__self__, "provisioning_state", provisioning_state)
if public_ip_id and not isinstance(public_ip_id, str):
raise TypeError("Expected argument 'public_ip_id' to be a str")
pulumi.set(__self__, "public_ip_id", public_ip_id)
if support and not isinstance(support, dict):
raise TypeError("Expected argument 'support' to be a dict")
pulumi.set(__self__, "support", support)
if tags and not isinstance(tags, dict):
raise TypeError("Expected argument 'tags' to be a dict")
pulumi.set(__self__, "tags", tags)
if type and not isinstance(type, str):
raise TypeError("Expected argument 'type' to be a str")
pulumi.set(__self__, "type", type)
if unique_identifier and not isinstance(unique_identifier, str):
raise TypeError("Expected argument 'unique_identifier' to be a str")
pulumi.set(__self__, "unique_identifier", unique_identifier)
if vault_name and not isinstance(vault_name, str):
raise TypeError("Expected argument 'vault_name' to be a str")
pulumi.set(__self__, "vault_name", vault_name)
if vm_creation_resource_group and not isinstance(vm_creation_resource_group, str):
raise TypeError("Expected argument 'vm_creation_resource_group' to be a str")
pulumi.set(__self__, "vm_creation_resource_group", vm_creation_resource_group)
@property
@pulumi.getter
def announcement(self) -> Optional['outputs.LabAnnouncementPropertiesResponse']:
"""
The properties of any lab announcement associated with this lab
"""
return pulumi.get(self, "announcement")
@property
@pulumi.getter(name="artifactsStorageAccount")
def artifacts_storage_account(self) -> str:
"""
The lab's artifact storage account.
"""
return pulumi.get(self, "artifacts_storage_account")
@property
@pulumi.getter(name="createdDate")
def created_date(self) -> str:
"""
The creation date of the lab.
"""
return pulumi.get(self, "created_date")
@property
@pulumi.getter(name="defaultPremiumStorageAccount")
def default_premium_storage_account(self) -> str:
"""
The lab's default premium storage account.
"""
return pulumi.get(self, "default_premium_storage_account")
@property
@pulumi.getter(name="defaultStorageAccount")
def default_storage_account(self) -> str:
"""
The lab's default storage account.
"""
return pulumi.get(self, "default_storage_account")
@property
@pulumi.getter(name="environmentPermission")
def environment_permission(self) -> Optional[str]:
"""
The access rights to be granted to the user when provisioning an environment
"""
return pulumi.get(self, "environment_permission")
@property
@pulumi.getter(name="extendedProperties")
def extended_properties(self) -> Optional[Mapping[str, str]]:
"""
Extended properties of the lab used for experimental features
"""
return pulumi.get(self, "extended_properties")
@property
@pulumi.getter(name="labStorageType")
def lab_storage_type(self) -> Optional[str]:
"""
Type of storage used by the lab. It can be either Premium or Standard. Default is Premium.
"""
return pulumi.get(self, "lab_storage_type")
@property
@pulumi.getter(name="loadBalancerId")
def load_balancer_id(self) -> str:
"""
The load balancer used to for lab VMs that use shared IP address.
"""
return pulumi.get(self, "load_balancer_id")
@property
@pulumi.getter
def location(self) -> Optional[str]:
"""
The location of the resource.
"""
return pulumi.get(self, "location")
@property
@pulumi.getter(name="mandatoryArtifactsResourceIdsLinux")
def mandatory_artifacts_resource_ids_linux(self) -> Optional[Sequence[str]]:
"""
The ordered list of artifact resource IDs that should be applied on all Linux VM creations by default, prior to the artifacts specified by the user.
"""
return pulumi.get(self, "mandatory_artifacts_resource_ids_linux")
@property
@pulumi.getter(name="mandatoryArtifactsResourceIdsWindows")
def mandatory_artifacts_resource_ids_windows(self) -> Optional[Sequence[str]]:
"""
The ordered list of artifact resource IDs that should be applied on all Windows VM creations by default, prior to the artifacts specified by the user.
"""
return pulumi.get(self, "mandatory_artifacts_resource_ids_windows")
@property
@pulumi.getter
def name(self) -> str:
"""
The name of the resource.
"""
return pulumi.get(self, "name")
@property
@pulumi.getter(name="networkSecurityGroupId")
def network_security_group_id(self) -> str:
"""
The Network Security Group attached to the lab VMs Network interfaces to restrict open ports.
"""
return pulumi.get(self, "network_security_group_id")
@property
@pulumi.getter(name="premiumDataDiskStorageAccount")
def premium_data_disk_storage_account(self) -> str:
"""
The lab's premium data disk storage account.
"""
return pulumi.get(self, "premium_data_disk_storage_account")
@property
@pulumi.getter(name="premiumDataDisks")
def premium_data_disks(self) -> Optional[str]:
"""
The setting to enable usage of premium data disks.
When its value is 'Enabled', creation of standard or premium data disks is allowed.
When its value is 'Disabled', only creation of standard data disks is allowed.
"""
return pulumi.get(self, "premium_data_disks")
@property
@pulumi.getter(name="provisioningState")
def provisioning_state(self) -> str:
"""
The provisioning status of the resource.
"""
return pulumi.get(self, "provisioning_state")
@property
@pulumi.getter(name="publicIpId")
def public_ip_id(self) -> str:
"""
The public IP address for the lab's load balancer.
"""
return pulumi.get(self, "public_ip_id")
@property
@pulumi.getter
def support(self) -> Optional['outputs.LabSupportPropertiesResponse']:
"""
The properties of any lab support message associated with this lab
"""
return pulumi.get(self, "support")
@property
@pulumi.getter
def tags(self) -> Optional[Mapping[str, str]]:
"""
The tags of the resource.
"""
return pulumi.get(self, "tags")
@property
@pulumi.getter
def type(self) -> str:
"""
The type of the resource.
"""
return pulumi.get(self, "type")
@property
@pulumi.getter(name="uniqueIdentifier")
def unique_identifier(self) -> str:
"""
The unique immutable identifier of a resource (Guid).
"""
return pulumi.get(self, "unique_identifier")
@property
@pulumi.getter(name="vaultName")
def vault_name(self) -> str:
"""
The lab's Key vault.
"""
return pulumi.get(self, "vault_name")
@property
@pulumi.getter(name="vmCreationResourceGroup")
def vm_creation_resource_group(self) -> str:
"""
The resource group in which all new lab virtual machines will be created. To let DevTest Labs manage resource group creation, set this value to null.
"""
return pulumi.get(self, "vm_creation_resource_group")
class AwaitableGetLabResult(GetLabResult):
# pylint: disable=using-constant-test
def __await__(self):
if False:
yield self
return GetLabResult(
announcement=self.announcement,
artifacts_storage_account=self.artifacts_storage_account,
created_date=self.created_date,
default_premium_storage_account=self.default_premium_storage_account,
default_storage_account=self.default_storage_account,
environment_permission=self.environment_permission,
extended_properties=self.extended_properties,
lab_storage_type=self.lab_storage_type,
load_balancer_id=self.load_balancer_id,
location=self.location,
mandatory_artifacts_resource_ids_linux=self.mandatory_artifacts_resource_ids_linux,
mandatory_artifacts_resource_ids_windows=self.mandatory_artifacts_resource_ids_windows,
name=self.name,
network_security_group_id=self.network_security_group_id,
premium_data_disk_storage_account=self.premium_data_disk_storage_account,
premium_data_disks=self.premium_data_disks,
provisioning_state=self.provisioning_state,
public_ip_id=self.public_ip_id,
support=self.support,
tags=self.tags,
type=self.type,
unique_identifier=self.unique_identifier,
vault_name=self.vault_name,
vm_creation_resource_group=self.vm_creation_resource_group)
def get_lab(expand: Optional[str] = None,
name: Optional[str] = None,
resource_group_name: Optional[str] = None,
opts: Optional[pulumi.InvokeOptions] = None) -> AwaitableGetLabResult:
"""
Use this data source to access information about an existing resource.
:param str expand: Specify the $expand query. Example: 'properties($select=defaultStorageAccount)'
:param str name: The name of the lab.
:param str resource_group_name: The name of the resource group.
"""
__args__ = dict()
__args__['expand'] = expand
__args__['name'] = name
__args__['resourceGroupName'] = resource_group_name
if opts is None:
opts = pulumi.InvokeOptions()
if opts.version is None:
opts.version = _utilities.get_version()
__ret__ = pulumi.runtime.invoke('azure-nextgen:devtestlab/latest:getLab', __args__, opts=opts, typ=GetLabResult).value
return AwaitableGetLabResult(
announcement=__ret__.announcement,
artifacts_storage_account=__ret__.artifacts_storage_account,
created_date=__ret__.created_date,
default_premium_storage_account=__ret__.default_premium_storage_account,
default_storage_account=__ret__.default_storage_account,
environment_permission=__ret__.environment_permission,
extended_properties=__ret__.extended_properties,
lab_storage_type=__ret__.lab_storage_type,
load_balancer_id=__ret__.load_balancer_id,
location=__ret__.location,
mandatory_artifacts_resource_ids_linux=__ret__.mandatory_artifacts_resource_ids_linux,
mandatory_artifacts_resource_ids_windows=__ret__.mandatory_artifacts_resource_ids_windows,
name=__ret__.name,
network_security_group_id=__ret__.network_security_group_id,
premium_data_disk_storage_account=__ret__.premium_data_disk_storage_account,
premium_data_disks=__ret__.premium_data_disks,
provisioning_state=__ret__.provisioning_state,
public_ip_id=__ret__.public_ip_id,
support=__ret__.support,
tags=__ret__.tags,
type=__ret__.type,
unique_identifier=__ret__.unique_identifier,
vault_name=__ret__.vault_name,
vm_creation_resource_group=__ret__.vm_creation_resource_group)
| [
"public@paulstack.co.uk"
] | public@paulstack.co.uk |
80e5198c86ed27d6c8417c69bd3be133fb25d9a9 | 87738f6f80f4fe03ac15d6de02b9b51f4e340c4b | /Part 3 - Classification/Section 10 - Support Vector Machine (SVM)/svm.py | d94e5cd0a3aee556681f4db83424c2c51038813a | [
"MIT"
] | permissive | Nikronic/Machine-Learning-Models | b4209dcc5f5a428da9447bc40fc93b9f0e171880 | 9fb48463ee2211eec800b2436699508f55d5ee28 | refs/heads/master | 2021-11-10T14:50:56.425443 | 2021-11-07T15:42:48 | 2021-11-07T15:42:48 | 142,788,604 | 34 | 13 | null | null | null | null | UTF-8 | Python | false | false | 3,824 | py | # -*- coding: utf-8 -*-
"""
Created on Fri Aug 17 23:26:31 2018
@author: Mohammad Doosti Lakhani
"""
# Imporing libraries
import numpy as np
import matplotlib.pyplot as plt
import pandas as pd
# Importing dataset
dataset = pd.read_csv('Social_Network_Ads.csv')
x = dataset.iloc[:,2:4].values
y = dataset.iloc[:,4].values
# Feature scaling
from sklearn.preprocessing import StandardScaler
standardscaler_x = StandardScaler()
standardscaler_x = standardscaler_x.fit(x)
x = standardscaler_x.transform(x)
# Splitting dataset into Train set and Test set
from sklearn.model_selection import train_test_split
x_train,x_test,y_train,y_test = train_test_split(x,y, train_size = 0.75 , random_state=0)
# Fitting the K-Nearest Neighbors model to the train set
from sklearn.svm import SVC
classifier = SVC(kernel='rbf', random_state=0)
classifier = classifier.fit(x_train,y_train)
""" Try to uncomment below code and see the visualization output (Try different kernels)"""
"""
classifier = SVC(kernel='poly', degree = 3, random_state=0)
classifier = classifier.fit(x_train,y_train)
classifier = SVC(kernel='linear', random_state=0) ## (Equals to SVR)
classifier = classifier.fit(x_train,y_train)
classifier = SVC(kernel='sigmoid', random_state=0)
classifier = classifier.fit(x_train,y_train)
"""
# Make the prediction on train set
y_train_pred = classifier.predict(x_train)
# Make the prediction on train set
y_test_pred = classifier.predict(x_test)
# Acurracy on test and train set
from sklearn.metrics import confusion_matrix
cm_train = confusion_matrix(y_train,y_train_pred)
cm_test = confusion_matrix(y_test,y_test_pred)
import os
import sys
scriptpath = "../../Tools" # functions of acc and CAP
# Add the directory containing your module to the Python path
sys.path.append(os.path.abspath(scriptpath))
import accuracy as ac
t_train,f_train,acc_train = ac.accuracy_on_cm(cm_train)
print('Train status = #{} True, #{} False, %{} Accuracy'.format(t_train,f_train,acc_train*100))
t_test,f_test,acc_test = ac.accuracy_on_cm(cm_test)
print('Test status = #{} True, #{} False, %{} Accuracy'.format(t_test,f_test,acc_test*100))
# Visualising the Training set results
from matplotlib.colors import ListedColormap
X_set, y_set = x_train, y_train
X1, X2 = np.meshgrid(np.arange(start = X_set[:, 0].min() - 1, stop = X_set[:, 0].max() + 1, step = 0.01),
np.arange(start = X_set[:, 1].min() - 1, stop = X_set[:, 1].max() + 1, step = 0.01))
plt.contourf(X1, X2, classifier.predict(np.array([X1.ravel(), X2.ravel()]).T).reshape(X1.shape),
alpha = 0.75, cmap = ListedColormap(('red', 'green')))
plt.xlim(X1.min(), X1.max())
plt.ylim(X2.min(), X2.max())
for i, j in enumerate(np.unique(y_set)):
plt.scatter(X_set[y_set == j, 0], X_set[y_set == j, 1],
c = ListedColormap(('red', 'green'))(i), label = j)
plt.title('SVM - rbf kernel (Training set)')
plt.xlabel('Age')
plt.ylabel('Estimated Salary')
plt.legend()
plt.show()
# Visualising the Test set results
from matplotlib.colors import ListedColormap
X_set, y_set = x_test, y_test
X1, X2 = np.meshgrid(np.arange(start = X_set[:, 0].min() - 1, stop = X_set[:, 0].max() + 1, step = 0.01),
np.arange(start = X_set[:, 1].min() - 1, stop = X_set[:, 1].max() + 1, step = 0.01))
plt.contourf(X1, X2, classifier.predict(np.array([X1.ravel(), X2.ravel()]).T).reshape(X1.shape),
alpha = 0.75, cmap = ListedColormap(('red', 'green')))
plt.xlim(X1.min(), X1.max())
plt.ylim(X2.min(), X2.max())
for i, j in enumerate(np.unique(y_set)):
plt.scatter(X_set[y_set == j, 0], X_set[y_set == j, 1],
c = ListedColormap(('red', 'green'))(i), label = j)
plt.title('SVM - rbf kernel (Test set)')
plt.xlabel('Age')
plt.ylabel('Estimated Salary')
plt.legend()
plt.show()
| [
"nikan.doosti@outlook.com"
] | nikan.doosti@outlook.com |
9889f501e5014b0b7e9eb35d9828310da06d2dac | 879b62155c70622d4100bf44c219f11288d90d76 | /peak_detection_matlab.py | 3302b3604bd5c0d46c5ca0f9fc18171ac27db9b3 | [] | no_license | praneelrathore/Ballistocardiography | c9e53d714659c91814f40ca59f24d2e127de3a93 | c8f8c2841f53bbacfb2179fd3aef44f3a58d3afa | refs/heads/master | 2021-11-11T17:12:11.816019 | 2021-11-02T16:41:41 | 2021-11-02T16:41:41 | 74,914,249 | 2 | 1 | null | null | null | null | UTF-8 | Python | false | false | 2,114 | py | def beat_location(frames_after_spline, window_size, cs_new, Fs):
k = 0
x_index = []#zeros()
beat_point = []#zeros()
index = 0
find_o = 0
window = 0
while 1:
if cs_new[find_o] == 0:
break
elif((cs_new[find_o] > 0 and cs_new[find_o + 1] < 0) or cs_new[find_o] < 0 and cs_new[find_o + 1] > 0):
break
else:
find_o = find_o + 1
for i in range(0, frames_after_spline+1, window_size):
if i == 0:
maxi = 0
lower = find_o
else:
maxi = 0
if k == 0:
lower = window_size
elif(window - index) < Fs / 5:
lower = x_index(k) + Fs / 2
if (lower >= frames_after_spline):
lower = frames_after_spline - 1
check = i
while check < lower:
if cs_new(check) > cs_new(check - 1) and cs_new(check) > cs_new(check + 1) and cs_new(check) > beat_point(k) and cs_new(check) > cs_new(check - 15) and cs_new(check) > cs_new(check + 15):
beat_point[k] = cs_new(check)
x_index[k] = check
else:
check = check + 1
else:
lower = i
flag = 0
window = i + window_size - 1
if window - frames_after_spline > Fs / 5:
break
elif window > frames_after_spline:
window = frames_after_spline
for j in range(lower-1, window+1):
if (j != frames_after_spline and j < frames_after_spline) and cs_new[j] > cs_new[j+1]:
if j != 0 and cs_new[j] > cs_new[j - 1] and cs_new[j] > 0:
if maxi < cs_new[j]:
flag = 1
maxi = cs_new[j]
index = j
if flag != 0:
x_index.append(index)
beat_point.append(cs_new[index])
#k = k + 1;
#x_index[k] = index;
#beat_point[k] = cs_new[index];
return beat_point, x_index
| [
"a.garg95@gmail.com"
] | a.garg95@gmail.com |
ae4cb13734a0740053a6c4093337ac9c7f2ab6d8 | de707c94c91f554d549e604737b72e6c86eb0755 | /supervised_learning/0x02-tensorflow/7-evaluate.py | 16e4666e5785a1670cb87f5a081e939092818dc2 | [] | no_license | ejonakodra/holbertonschool-machine_learning-1 | 885cf89c1737573228071e4dc8e26304f393bc30 | 8834b201ca84937365e4dcc0fac978656cdf5293 | refs/heads/main | 2023-07-10T09:11:01.298863 | 2021-08-11T03:43:59 | 2021-08-11T03:43:59 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,069 | py | #!/usr/bin/env python3
"""
Defines a function that evaluates output of
neural network classifier
"""
import tensorflow as tf
def evaluate(X, Y, save_path):
"""
Evaluates output of neural network
parameters:
X [numpy.ndarray]: contains the input data to evaluate
Y [numpy.ndarray]: contains the one-hot labels for X
save_path [string]: location to load the model from
returns:
the network's prediction, accuracy, and loss, respectively
"""
with tf.Session() as sess:
saver = tf.train.import_meta_graph(save_path + '.meta')
saver.restore(sess, save_path)
x = tf.get_collection('x')[0]
y = tf.get_collection('y')[0]
y_pred = tf.get_collection('y_pred')[0]
accuracy = tf.get_collection('accuracy')[0]
loss = tf.get_collection('loss')[0]
prediction = sess.run(y_pred, feed_dict={x: X, y: Y})
accuracy = sess.run(accuracy, feed_dict={x: X, y: Y})
loss = sess.run(loss, feed_dict={x: X, y: Y})
return (prediction, accuracy, loss)
| [
"eislek02@gmail.com"
] | eislek02@gmail.com |
0aced39f9e720a7bd36e2e19e7abdce7a403c575 | 6b2d62130252c0921715b22a626a2ce94ba91ee1 | /ITP125/Homework/Homework 3/hw03a.py | 1691b1ab48b6e8db263cbb850bf92b492a0d76e4 | [] | no_license | sthsher/USC_CECS | fba48c7b331f562a802958a9412ab2d5ee1f2ec3 | 776ca95cab2b1c132ee87e1fd8ce95c77315e452 | refs/heads/master | 2018-12-11T10:42:58.576051 | 2016-11-01T06:20:11 | 2016-11-01T06:20:11 | 24,713,040 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 404 | py | shopping_list = ["banana", "orange", "apple"]
stock = {
"banana": 6,
"apple": 0,
"orange": 32,
"pear": 15
}
prices = {
"banana": 4,
"apple": 2,
"orange": 1.5,
"pear": 3
}
# Write your code below!
def compute_bill(food):
total = 0
for item in food:
if (stock[item] > 0):
total += prices[item]
stock[item] -= 1
return total | [
"stephen.sher.94@gmail.com"
] | stephen.sher.94@gmail.com |
4f8cdbd5a7b33f05ef45e53e15cec978c09c4e1f | a86acf28472700008261fb3709134483b4d4c7e2 | /myutils/seq2seq_peeky.py | 39147587ca4e866f28e07cc94f092969ca0bb185 | [] | no_license | s1len7/ai-rnn | 5a0ab40e7aa0371f41131e99371991c3aeecfa98 | e3f0a07f1a5e4681b2209276b5f3d6a5d4901f1d | refs/heads/master | 2022-11-30T13:55:41.267065 | 2020-07-29T11:19:18 | 2020-07-29T11:19:18 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,012 | py | # -*- coding: utf-8 -*-
"""
Created on Sun Mar 1 11:24:24 2020
@author: shkim
"""
"""
## seq2seq 성능 개선
* 1) 입력 데이터 반전(Reverse)
* 2) 엿보기(Peeky)
"""
"""
## seq2seq 성능 개선 : 엿보기(Peeky)
### base seq2seq 모델에서의 동작
* Encoder는 입력문장(문제문장)을 고정길이 벡터 h로 변환함
* 이때 h안에는 Decoder에 필요한 정보가 모두 담겨 있음
* 즉 h가 Decoder에 있어서 유일한 정보인 셈임
* 최초 시각의 LSTM 계층만이 벡터 h를 이용함 --> 이 중요한 h 정보를 더 활용할 수 없을까?
### 개선된 seq2seq 모델 : 엿보기(Peeky) 모델
* 중요한 정보가 담긴 Encoder의 출력 h를 Decoder의 다른 계층에도 전달해 주는 것
* Encoder의 출력 h를 모든 시각의 LSTM 계층과 Affine 계층에 전해줌 --> 집단 지성
* LSTM 계층과 Affine 계층에 입력되는 벡터가 2개씩 됨 --> concatenate 됨
"""
#%%
"""
## 개선된 seq2seq 모델 : 엿보기(Peeky) 모델 구현
"""
import numpy as np
import sys
sys.path.append('..')
from myutils.time_layers import TimeEmbedding, TimeLSTM, TimeAffine, TimeSoftmaxWithLoss
from myutils.seq2seq import Seq2seq, Encoder
#%%
class DecoderPeeky:
def __init__(self, vocab_size, wordvec_size, hideen_size):
V, D, H = vocab_size, wordvec_size, hideen_size
rn = np.random.randn
embed_W = (rn(V, D) / 100).astype('f')
lstm_Wx = (rn(H+D, 4*H) / np.sqrt(H+D)).astype('f')
lstm_Wh = (rn(H, 4*H) / np.sqrt(H)).astype('f')
lstm_b = np.zeros(4*H).astype('f')
affine_W = (rn(H+H, V) / np.sqrt(H+H)).astype('f')
affine_b = np.zeros(V).astype('f')
self.embed = TimeEmbedding(embed_W)
self.lstm = TimeLSTM(lstm_Wx, lstm_Wh, lstm_b, stateful=True)
self.affine = TimeAffine(affine_W, affine_b)
self.params, self.grads = [], []
for layer in (self.embed, self.lstm, self.affine):
self.params += layer.params
self.grads += layer.grads
self.cache = None
def forward(self, xs, h):
N, T = xs.shape
N, H = h.shape
self.lstm.set_state(h)
out = self.embed.forward(xs)
hs = np.repeat(h, T, axis=0).reshape(N, T, H)
out = np.concatenate((hs, out), axis=2)
out = self.lstm.forward(out)
out = np.concatenate((hs, out), axis=2)
score = self.affine.forward(out)
self.cache = H
return score
def backward(self, dscore):
H = self.cache
dout = self.affine.backward(dscore)
dout, dhs0 = dout[:, :, H:], dout[:, :, :H]
dout = self.lstm.backward(dout)
dembed, dhs1 = dout[:, :, H:], dout[:, :, :H]
self.embed.backward(dembed)
dhs = dhs0 + dhs1
dh = self.lstm.dh + np.sum(dhs, axis=1)
return dh
def generate(self, h, start_id, sample_size):
sampled = []
char_id = start_id
self.lstm.set_state(h)
H = h.shape[1]
peeky_h = h.reshape(1, 1, H)
for _ in range(sample_size):
x = np.array([char_id]).reshape((1, 1))
out = self.embed.forward(x)
out = np.concatenate((peeky_h, out), axis=2)
out = self.lstm.forward(out)
out = np.concatenate((peeky_h, out), axis=2)
score = self.affine.forward(out)
char_id = np.argmax(score.flatten())
sampled.append(char_id)
return sampled
#%%
class Seq2seqPeeky(Seq2seq):
def __init__(self, vocab_size, wordvec_size, hidden_size):
V, D, H = vocab_size, wordvec_size, hidden_size
self.encoder = Encoder(V, D, H)
self.decoder = DecoderPeeky(V, D, H)
self.softmax = TimeSoftmaxWithLoss()
self.params = self.encoder.params + self.decoder.params
self.grads = self.encoder.grads + self.decoder.grads
#%%
| [
"shkim.hi@gmail.com"
] | shkim.hi@gmail.com |
3b7f9e6dbe9c7e658110923f1a4756af7ddbc9ba | 53fab060fa262e5d5026e0807d93c75fb81e67b9 | /backup/user_274/ch134_2020_04_01_11_05_19_507472.py | ead9069b851dccdce2014878c391a3cdbe73018b | [] | no_license | gabriellaec/desoft-analise-exercicios | b77c6999424c5ce7e44086a12589a0ad43d6adca | 01940ab0897aa6005764fc220b900e4d6161d36b | refs/heads/main | 2023-01-31T17:19:42.050628 | 2020-12-16T05:21:31 | 2020-12-16T05:21:31 | 306,735,108 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 174 | py | def verifica_quadrado_perfeito(n):
m=n
i=2
while m > -1:
m=m-i
i=i+2
if m**2 == n:
return True
else:
return False | [
"you@example.com"
] | you@example.com |
2b010340767a1cea96729c9cee25bcdfd1e4b981 | c6b53079177e99790651980910358f9643eff009 | /blog/migrations/0009_blogpost_user.py | c466dc61885fc1ccf07ec3e976e2e2352c4eb528 | [] | no_license | ravi19999/Try-Django-2.2 | 09bffcf39806fce4fdbb4ff59f9fc5934ba9991a | f0379445bdb7fc07c14bcbf7514438ede3700223 | refs/heads/master | 2023-02-07T02:15:43.244332 | 2020-12-19T14:26:13 | 2020-12-19T14:26:13 | 303,755,405 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 595 | py | # Generated by Django 2.2 on 2019-07-28 17:51
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
('blog', '0008_auto_20190727_1119'),
]
operations = [
migrations.AddField(
model_name='blogpost',
name='user',
field=models.ForeignKey(default=1, null=True, on_delete=django.db.models.deletion.SET_NULL, to=settings.AUTH_USER_MODEL),
),
]
| [
"dahalravindra@gmail.com"
] | dahalravindra@gmail.com |
81264f2bcadaa766a81e3a63ef481439ed76116f | e20ed90b9be7a0bcdc1603929d65b2375a224bf6 | /generated-libraries/python/netapp/fpolicy/event_name.py | 476ecd9c31bd1a9cacb1652502a4f667427125da | [
"MIT"
] | permissive | radekg/netapp-ontap-lib-gen | 530ec3248cff5ead37dc2aa47ced300b7585361b | 6445ebb071ec147ea82a486fbe9f094c56c5c40d | refs/heads/master | 2016-09-06T17:41:23.263133 | 2015-01-14T17:40:46 | 2015-01-14T17:40:46 | 29,256,898 | 2 | 0 | null | null | null | null | UTF-8 | Python | false | false | 142 | py | class EventName(basestring):
"""
Event name
"""
@staticmethod
def get_api_name():
return "event-name"
| [
"radek@gruchalski.com"
] | radek@gruchalski.com |
b054c26aafaec13538eb27a9368e36aabd141663 | 5da914bac3349b7ddbc31b17e43e9da99ae8f462 | /Ultimo Backup Server/node_modules/websocket/build/config.gypi | 46dc3b48213b03d7ddb4dbb87ef42c1eb094d654 | [] | no_license | marox360/Cemin-Marogna_workspace | 0bcd9fa5f508b250b0b57b4dc21254a9f24a64d1 | 63c630d95ca523f2e079b7d3831a6b45daca55bf | refs/heads/master | 2021-03-12T23:16:34.845956 | 2014-11-16T11:54:25 | 2014-11-16T11:54:25 | 10,483,752 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 296 | gypi | # Do not edit. File was generated by node-gyp's "configure" step
{
"target_defaults": {
"cflags": [],
"defines": [],
"include_dirs": [],
"libraries": [],
"default_configuration": "Release"
},
"variables": {
"target_arch": "arm",
"target_version": "0.6.17"
}
}
| [
"lmarox@gmail.com"
] | lmarox@gmail.com |
77e13c60ab887ef65af5d208fbcad6ac63b78f87 | f067b46c0bd8bf4fbc2471c42c4a74cb08359bd5 | /server/config/settings/components/thumb.py | 6b24b0bc3fdca3195cad3f422c9b6577525857e0 | [] | no_license | bopo/project-template | ffbf3516c9f486fadb46a767688cb26badda6a3d | 0eedd18c236b66516e543750673934f4932555ca | refs/heads/develop | 2020-03-23T01:57:29.471378 | 2018-07-17T05:38:48 | 2018-07-17T05:38:48 | 140,947,688 | 0 | 1 | null | 2022-04-21T04:40:56 | 2018-07-14T13:18:49 | Python | UTF-8 | Python | false | false | 1,330 | py | # -*- coding: utf-8 -*-
# INSTALLED_APPS += [
# "django_thumbor",
# ]
# INSTALLED_APPS += ('easy_thumbnails',)
# THUMBNAIL_ALIASES = {
# '': {
# 'avatar': {'size': (50, 50), 'crop': True},
# },
# }
# THUMB_LIST = '500x500'
# THUMB_DETAIL = '800x800'
# The host serving the thumbor resized images
THUMBOR_SERVER = 'http://localhost:8888'
# The prefix for the host serving the original images
# This must be a resolvable address to allow thumbor to reach the images
THUMBOR_MEDIA_URL = 'http://localhost:8888/media'
# If you want the static to be handled by django thumbor
# default as False, set True to handle it if you host your statics
THUMBOR_STATIC_ENABLED = False
# The prefix for the host serving the original static images
# this must be a resolvable address to allow thumbor to reach the images
THUMBOR_STATIC_URL = 'http://localhost:8888/static'
# The same security key used in the thumbor service to
# match the URL construction
THUMBOR_SECURITY_KEY = 'MY_SECURE_KEY'
# Default arguments passed to the `generate_url` helper or
# the `thumbor_url` templatetag
THUMBOR_ARGUMENTS = {}
# An alias represents a named set of arguments to the generate_url function
# or thumbor_url template tag. Use it to share general thumbnail
# configurations without repeating yourself.
THUMBOR_ALIASES = {}
| [
"travis@travis-ci.org"
] | travis@travis-ci.org |
c74aea42a9ebf71293d70dbcefafdea417e21e02 | f66b0fda3a16eb7ceb8a7ceb63dcbc373c4ece2d | /smart_QC/libs/json_field/fields.py | 7fcada3804ff579ad916c9cef24d6cb98f62d8c2 | [] | no_license | zcyuefan/smart_QC | 9c82dd7b8a25bfcd0231195b1a797e3f491f3272 | fc8aaec48203065f09ffde7f318c3e39680092ba | refs/heads/master | 2021-01-19T04:55:40.300511 | 2017-09-13T02:31:48 | 2017-09-13T02:31:48 | 81,196,835 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 8,430 | py | from __future__ import unicode_literals
from .utils import is_aware
from .forms import JSONFormField
try:
import json
except ImportError: # python < 2.6
from django.utils import simplejson as json
from django.db import models
from django.core import exceptions
from django.utils.translation import ugettext_lazy as _
from django.core.exceptions import ImproperlyConfigured
import re
import decimal
import datetime
import six
try:
from dateutil import parser as date_parser
except ImportError:
raise ImproperlyConfigured('The "dateutil" library is required and was not found.')
try:
JSON_DECODE_ERROR = json.JSONDecodeError # simplejson
except AttributeError:
JSON_DECODE_ERROR = ValueError # other
TIME_FMT = r'\d{2}:\d{2}:\d{2}(\.\d+)?'
DATE_FMT = r'\d{4}-\d{2}-\d{2}'
TIMEZONE_FMT = r'(\+|\-)\d{2}:\d{2}'
TIME_RE = re.compile(r'^(%s)$' % TIME_FMT)
DATE_RE = re.compile(r'^(%s)$' % DATE_FMT)
DATETIME_RE = re.compile(r'^(%s)T(%s)(%s)?$' % (DATE_FMT, TIME_FMT, TIMEZONE_FMT))
class JSONEncoder(json.JSONEncoder):
"""
JSONEncoder subclass that knows how to encode date/time and decimal types.
"""
def default(self, o):
# See "Date Time String Format" in the ECMA-262 specification.
if isinstance(o, datetime.datetime):
r = o.isoformat()
if o.microsecond:
r = r[:23] + r[26:]
if r.endswith('+00:00'):
r = r[:-6] + 'Z'
return r
elif isinstance(o, datetime.date):
return o.isoformat()
elif isinstance(o, datetime.time):
if is_aware(o):
raise ValueError("JSON can't represent timezone-aware times.")
r = o.isoformat()
if o.microsecond:
r = r[:12]
return r
elif isinstance(o, decimal.Decimal):
return str(o)
else:
return super(JSONEncoder, self).default(o)
class JSONDecoder(json.JSONDecoder):
""" Recursive JSON to Python deserialization. """
_recursable_types = ([str] if six.PY3 else [str, unicode]) + [list, dict]
def _is_recursive(self, obj):
return type(obj) in JSONDecoder._recursable_types
def decode(self, obj, *args, **kwargs):
if not kwargs.get('recurse', False):
obj = super(JSONDecoder, self).decode(obj, *args, **kwargs)
if isinstance(obj, list):
for i in six.moves.xrange(len(obj)):
item = obj[i]
if self._is_recursive(item):
obj[i] = self.decode(item, recurse=True)
elif isinstance(obj, dict):
for key, value in obj.items():
if self._is_recursive(value):
obj[key] = self.decode(value, recurse=True)
elif isinstance(obj, six.string_types):
if TIME_RE.match(obj):
try:
return date_parser.parse(obj).time()
except ValueError:
pass
if DATE_RE.match(obj):
try:
return date_parser.parse(obj).date()
except ValueError:
pass
if DATETIME_RE.match(obj):
try:
return date_parser.parse(obj)
except ValueError:
pass
return obj
class Creator(object):
"""
Taken from django.db.models.fields.subclassing.
"""
_state_key = '_json_field_state'
def __init__(self, field, lazy):
self.field = field
self.lazy = lazy
def __get__(self, obj, type=None):
if obj is None:
return self
if self.lazy:
state = getattr(obj, self._state_key, None)
if state is None:
state = {}
setattr(obj, self._state_key, state)
if state.get(self.field.name, False):
return obj.__dict__[self.field.name]
value = self.field.to_python(obj.__dict__[self.field.name])
obj.__dict__[self.field.name] = value
state[self.field.name] = True
else:
value = obj.__dict__[self.field.name]
return value
def __set__(self, obj, value):
obj.__dict__[self.field.name] = value if self.lazy else self.field.to_python(value)
class JSONField(models.TextField):
""" Stores and loads valid JSON objects. """
description = 'JSON object'
def __init__(self, *args, **kwargs):
self.default_error_messages = {
'invalid': _('Enter a valid JSON object')
}
self._db_type = kwargs.pop('db_type', None)
self.evaluate_formfield = kwargs.pop('evaluate_formfield', False)
self.lazy = kwargs.pop('lazy', True)
encoder = kwargs.pop('encoder', JSONEncoder)
decoder = kwargs.pop('decoder', JSONDecoder)
encoder_kwargs = kwargs.pop('encoder_kwargs', {})
decoder_kwargs = kwargs.pop('decoder_kwargs', {})
if not encoder_kwargs and encoder:
encoder_kwargs.update({'cls':encoder})
if not decoder_kwargs and decoder:
decoder_kwargs.update({'cls':decoder, 'parse_float':decimal.Decimal})
self.encoder_kwargs = encoder_kwargs
self.decoder_kwargs = decoder_kwargs
self.ignore_error = kwargs.pop('ignore_error', False)
kwargs['default'] = kwargs.get('default', 'null')
kwargs['help_text'] = kwargs.get('help_text', self.default_error_messages['invalid'])
super(JSONField, self).__init__(*args, **kwargs)
def db_type(self, *args, **kwargs):
if self._db_type:
return self._db_type
return super(JSONField, self).db_type(*args, **kwargs)
def to_python(self, value):
if value is None: # allow blank objects
return None
if isinstance(value, six.string_types):
try:
value = json.loads(value, **self.decoder_kwargs)
except JSON_DECODE_ERROR:
pass
return value
def get_db_prep_value(self, value, *args, **kwargs):
if self.null and value is None and not kwargs.get('force'):
return None
a=isinstance(value, six.string_types)
if isinstance(value, six.string_types) and self.ignore_error:
return value
return json.dumps(value, **self.encoder_kwargs)
def value_to_string(self, obj):
return self.get_db_prep_value(self._get_val_from_obj(obj))
# def value_to_string(self, obj):
# value = self.value_from_object(obj)
# return self.get_prep_value(value)
def value_from_object(self, obj):
raw_value = super(JSONField, self).value_from_object(obj)
a=isinstance(raw_value, six.string_types)
if isinstance(raw_value, six.string_types) and self.ignore_error:
return raw_value
return json.dumps(raw_value, **self.encoder_kwargs)
# return json.dumps(super(JSONField, self).value_from_object(obj), **self.encoder_kwargs)
def formfield(self, **kwargs):
defaults = {
'form_class': kwargs.get('form_class', JSONFormField),
'evaluate': self.evaluate_formfield,
'encoder_kwargs': self.encoder_kwargs,
'decoder_kwargs': self.decoder_kwargs,
'ignore_error': self.ignore_error
}
defaults.update(kwargs)
return super(JSONField, self).formfield(**defaults)
def contribute_to_class(self, cls, name):
super(JSONField, self).contribute_to_class(cls, name)
def get_json(model_instance):
return self.get_db_prep_value(getattr(model_instance, self.attname, None), force=True)
setattr(cls, 'get_%s_json' % self.name, get_json)
def set_json(model_instance, value):
return setattr(model_instance, self.attname, self.to_python(value))
setattr(cls, 'set_%s_json' % self.name, set_json)
setattr(cls, name, Creator(self, lazy=self.lazy)) # deferred deserialization
try:
# add support for South migrations
from south.modelsinspector import add_introspection_rules
rules = [
(
(JSONField,),
[],
{
'db_type': ['_db_type', {'default': None}]
}
)
]
add_introspection_rules(rules, ['^json_field\.fields\.JSONField'])
except ImportError:
pass
| [
"zcyuefan@126.com"
] | zcyuefan@126.com |
3c845b7e58c6b79f531cda02a940c22b1d703558 | ddbeb902aac1095e704f8dbd785c80b5054fd61c | /rango/models.py | a302b7c2f19e71a49547c30ee73ebd2278d7e3d7 | [] | no_license | Zhang9494/tango_with_django_project | ba355fcbd2cf201a6e849b31edbc615fd1472595 | 75e77cf6bdf1a7fcef1f02144ce9eaccc4aedfab | refs/heads/master | 2020-04-22T02:02:56.283865 | 2019-02-12T15:53:27 | 2019-02-12T15:53:27 | 170,034,826 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 605 | py | from django.db import models
# Create your models here.
class Category(models.Model):
name = models.CharField(max_length=128, unique=True)
views = models.IntegerField(default=0)
likes = models.IntegerField(default=0)
class Meta:
verbose_name_plural = 'Category'
verbose_name = 'Category'
def __str__(self):
return self.name
class Page(models.Model):
category = models.ForeignKey(Category)
title = models.CharField(max_length=128)
url = models.URLField()
views = models.IntegerField(default=0)
def __str__(self):
return self.title
| [
"2410728Z@student.gla.ac.uk"
] | 2410728Z@student.gla.ac.uk |
0caed19093b2bb8c8baa715f300078ce5af57891 | bd5b5363f4924ded1f18a0511a9c8a5b6183417a | /python/gopher/menu.py | 1bbc5f5bbf04f1cd2eac1df88bbb76aa20eca0d2 | [
"MIT"
] | permissive | blabber/harbour-gophish | bd05970770a6704ae2722d69ad9e5fd57a9c55de | f864c02b4eec3b0396dc50a37c46ac1326749407 | refs/heads/master | 2021-08-24T01:16:54.144364 | 2017-12-04T12:12:12 | 2017-12-04T12:12:12 | 109,180,940 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,046 | py | """Contains functions and classes for the interaction with gopher menus and its
items.
"""
class MenuItem:
"""Represents an item in a gopher menu."""
def __init__(self, type, user_name, selector, host, port):
"""Initializes a new MenuItem instance.
The parameters correspond to the menu item components described in
RFC1436.
"""
self.type = type
self.user_name = user_name
self.selector = selector
self.host = host
self.port = port
def url(self):
"""Returns a URL locating the menu item."""
return str(self)
def __eq__(self, other):
"""Returns True if the ManuItem other represents the same menu item."""
return self.__dict__ == other.__dict__
def __str__(self):
"""Returns a human readable representation of the MenuItem.
At the moment, this method returns the gopher URL locating the menu
item.
"""
return 'gopher://{host}:{port}/{type}{selector}'.format(
**self.__dict__)
def __repr__(self):
"""Returns a string representation of the MenuItem."""
return ("MenuItem('{type}', '{user_name}', '{selector}', '{host}', "
"'{port}')").format(**self.__dict__)
def menuitem_from_raw_line(line):
"""Reads a line from a raw gopher menu and returns the corresponding
MenuItem instance.
As a special case, lines containing no '\t' seperators are interpreted as
informational messages (item type 'i').
"""
p = line.split('\t')
if len(p) == 1:
return MenuItem('i', p[0], '', 'fake', '70')
return MenuItem(p[0][0], p[0][1:], p[1], p[2], p[3])
def read_menu(request):
"""Reads a gopher menu and returns a list of MenuItems.
The menu that should be read is identified by request, a Request instance.
"""
items = []
for line in request.get_text_data().splitlines():
if line == '.':
break
items.append(menuitem_from_raw_line(line))
return items
| [
"tobias.rehbein@web.de"
] | tobias.rehbein@web.de |
368933543f3030bfc38b32795b89b4bccf0c2b47 | d8c1419eba8aeec8c203e819aae46475b744a66f | /archive/main.py | e6f339fa6539de2f2ff591d9c551fbb00f096b86 | [] | no_license | StefenYin/yeadon | a66aa1808ef662c76cd5d96db0f74cd25e3abcc7 | 03ae2c5881795e44f1890818fcb3530ba3c6feac | refs/heads/master | 2021-01-17T08:50:12.313236 | 2012-12-18T19:54:52 | 2012-12-18T19:54:52 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 698 | py | import stadium as stad
#import segment
import human as hum
import matplotlib.pyplot as mpl
from mpl_toolkits.mplot3d import Axes3D
import numpy as np
import data
import densities
# INPUTS ARE 95 MEASUREMENTS, DENSITIES, AND ORIENTATION ANGLES
# read input file of 95 measurements
# create solid objects
# create segment objects
# create human object
# plot human, no angles
# read in angles file
# plot human, with joint angles
# plot human conforming to a bicycle
# SECOND ITERATION: MOVE FROM FILE INPUTS (FOR ANGLES ONLY) TO QT GUI
externalangles = np.zeros( 3 )
externalangles[0] = 0
jointangles = np.zeros( 18 )
print "Creating human object."
H = hum.human(externalangles)
H.draw()
| [
"cld72@cornell.edu"
] | cld72@cornell.edu |
14b6673a73fd4152a4af7be21d6eb6e4233c7f7e | 944401a6292baa2d23b9738898e0b0cb199d0795 | /color_quantization/octree/img_quality_assessment(IQA)/psnr/rgb_cs/rgb_psnr_sky.py | cdd42100b4527e977262e59e0ed94e2810f09ba1 | [] | no_license | sunnyweilai/Finding-Theme-Color-Palettes | cc84c93ce58abdd1802431c41bd59181d7a4f75b | 4c38b112f5c40b43d6ec126e415b609c7fdc1f39 | refs/heads/master | 2022-12-21T09:41:31.187411 | 2019-04-30T14:50:17 | 2019-04-30T14:50:17 | 184,273,925 | 1 | 0 | null | 2022-12-07T03:46:55 | 2019-04-30T14:09:52 | Python | UTF-8 | Python | false | false | 1,028 | py | """
image quality assessment (IQA) of the quantized images and the original image in RGB color space
----- method: PSNR
----- version 1.0 ("skimage" library)
----- http://scikit-image.org/docs/dev/api/skimage.measure.html#skimage.measure.compare_psnr
"""
import numpy as np
from PIL import Image
import glob
import csv
import skimage.measure as skm
# ---- obtain the original and quantized images
temp_img = np.array(Image.open('../../../../img/sky.jpg'))
quantized_img_path_list = []
quantized_img_path_list = glob.glob(r'../../../img/sky/rgb_cs/quantized_img/*.png')
quantized_img_path_list.sort()
# ---- compute PSNR
score_list = []
for i in quantized_img_path_list:
quantized_img = np.array(Image.open(i))
score = skm.compare_psnr(temp_img, quantized_img)
score_list.append(score)
# print(score_list)
# ---- save psnr score to csv file
csvfile = "sky_psnr.csv"
with open(csvfile, "w") as output:
writer = csv.writer(output, lineterminator='\n')
for val in score_list:
writer.writerow([val]) | [
"wnn2260@gmail.com"
] | wnn2260@gmail.com |
98185120c68f6d1b952e23770e01e5e6897e04b1 | ea198bd6cf9b6c7cfa758b47807cf76279f4b049 | /Safety/users/migrations/0003_remove_user_radius_scan.py | c939b39ffb6f313db6133ca0f8b88484a4f4e128 | [] | no_license | hxt365/Safety | 2294e92e48ff1598a8a878e998f28fdf7f1eca8f | 0267337a5b4191e2b4c6fcc96c6493f3bcd47db1 | refs/heads/master | 2023-03-24T21:42:47.150843 | 2021-03-23T08:01:04 | 2021-03-23T08:01:04 | 334,204,269 | 3 | 0 | null | null | null | null | UTF-8 | Python | false | false | 329 | py | # Generated by Django 3.1.4 on 2020-12-04 23:46
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('users', '0002_auto_20201204_0111'),
]
operations = [
migrations.RemoveField(
model_name='user',
name='radius_scan',
),
]
| [
"hxt365@gmail.com"
] | hxt365@gmail.com |
0393192c5cd8c5fc21ce700e982b2dd3237f6f73 | 2ecfe34de36a0090154fb3a4fcb5785ce974eb09 | /GUI/RemoteBrowser.py | aa382124744e096fb2f815875bc75bb72676cc0a | [] | no_license | javiunzu/conan-gui | f362d39e14f1c5fba4ddcea93b7f14c8ea828116 | 681337f50abc6062cc504b7149dfcd1d90cbf635 | refs/heads/master | 2020-03-27T00:25:54.661918 | 2018-08-21T21:46:11 | 2018-08-21T21:46:11 | 145,622,426 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,568 | py | #!/usr/bin/python3
# -*- coding: utf-8 -*-
from PyQt5.QtWidgets import (QApplication, QMainWindow, QAction, qApp, QTabWidget, QSplitter)
from PyQt5.QtCore import Qt
from PyQt5.QtGui import QIcon
import json
import ConanCommander
#import MenuBar
from GUI import NavigationPanel
from GUI import PackagePanel
from GUI import GraphView
class RemoteBrowser(QMainWindow):
def __init__(self, cache_file):
super().__init__()
with open(cache_file) as fp:
self.__cache = json.load(fp)
self.commander = ConanCommander.ConanCommander()
self.init_menu_bar()
self.tabs = QTabWidget()
self.navigation = NavigationPanel.TreeView(self.cache)
self.details = PackagePanel.TreeView()
self.navigation.clicked.connect(self.onItemClicked)
self.graph = GraphView.GraphView()
self.tabs.addTab(self.details, "Details")
self.tabs.addTab(self.graph, "Graph")
self.splitter = QSplitter(Qt.Horizontal)
self.splitter.addWidget(self.navigation)
self.splitter.addWidget(self.tabs)
self.setCentralWidget(self.splitter)
self.statusBar().showMessage('Ready')
self.setGeometry(800, 600, 800, 600)
self.setWindowTitle('Remote Browser')
self.setWindowIcon(QIcon('web.png'))
self.show()
@property
def cache(self):
return self.__cache
def onItemClicked(self, index):
item = self.navigation.selectedIndexes()[0]
search = item.model().itemFromIndex(index)
result = self.commander.package_info(search.text(), search.parent().text())
self.details.populate(result, clear=True)
table = self.commander.package_table(search.text(), search.parent().text())
self.graph.load(table)
self.graph.show()
def init_menu_bar(self):
menubar = self.menuBar()
fileMenu = menubar.addMenu('&File')
aboutMenu = menubar.addMenu('&About')
remotes = QAction('Manage &Remotes', self)
remotes.setShortcut("Ctrl+Shift+R")
remotes.setStatusTip("Open the remote management window.")
fileMenu.addAction(remotes)
exitAct = QAction(QIcon('exit.png'), '&Exit', self)
exitAct.setShortcut('Ctrl+Q')
exitAct.setStatusTip('Exit application')
exitAct.triggered.connect(qApp.quit)
fileMenu.addAction(exitAct)
if __name__ == "__main__":
""" Do some trivial tests."""
import sys
app = QApplication(sys.argv)
browser = RemoteBrowser("../cache.json")
sys.exit(app.exec_()) | [
"javiunzu@gmail.com"
] | javiunzu@gmail.com |
08a31863c9164c94e5f21abf17a85789d3e0e5c6 | dfdd59d51c13adb6a9dde46fa1b184dde016bcad | /model/load_data.py | d2c1f7c14e41991a79e1aebec477dd14269da80e | [
"Apache-2.0"
] | permissive | googleinterns/cl_analysis | bf0e64e79d704e9fb7bba44751efffd6084fe80e | 686cf49cd57ce61cba3cc11f0574b2a2cec596be | refs/heads/master | 2022-12-16T03:56:34.473083 | 2020-09-26T06:44:12 | 2020-09-26T06:44:12 | 277,630,242 | 5 | 3 | Apache-2.0 | 2020-09-26T00:01:06 | 2020-07-06T19:26:14 | Python | UTF-8 | Python | false | false | 14,091 | py | # Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import pandas as pd
from collections import defaultdict
from datetime import datetime, timedelta
from copy import deepcopy
import os
from model.constants import *
class CLData:
"""Class that holds a piece of datum of a CL.
This class holds the pull request level features, file level features,
and the lable indicating whether the CL is reverted.
Attributes:
pr_level_features: A list of pull request level features.
file_level_features: A dict of file level features.
reverted: A boolean variable indicating the CL reversion.
"""
def __init__(self):
"""
Init the CL data.
"""
self.pr_level_features = None
self.file_level_features = {}
self.reverted = False
class DataLoader:
"""
This class helps load the whole dataset either from the local extracted
feature csv file or from the local saved txt file.
Attributes:
repos: A list of repo names.
pr_columns: A list of pull request level feature names.
file_columns: A list of file level feature names.
"""
def __init__(self, repos):
"""
Init DataLoader
Args:
repos: A list of repo names.
"""
self.repos = repos
self.pr_columns = COMMON_PR_LEVEL_FEATURES + EXTRA_PR_LEVEL_FEATURES
self.file_columns = COMMON_FILE_LEVEL_FEATURES + \
EXTRA_FILE_LEVEL_FEATURES
@staticmethod
def _count_check_run_passed(lst):
"""
Count the total number of passed check runs.
Args:
lst: A list of 'passed', 'failed', 'none'
Returns:
A integer indicating the total number of passed check runs.
"""
if pd.isna(lst):
return 0
num_passed = 0
for check_run_result in eval(lst):
if check_run_result == 'passed':
num_passed += 1
return num_passed
@staticmethod
def _count_check_run_failed(lst):
"""
Count the total number of failed check runs.
Args:
lst: A list of 'passed', 'failed', 'none'
Returns:
A integer indicating the total number of failed check runs.
"""
if pd.isna(lst):
return 0
num_failed = 0
for check_run_result in eval(lst):
if check_run_result == 'failed':
num_failed += 1
return num_failed
def _get_pr_level_signals(self, repo):
"""
Load the pull request level signals for input repo.
Args:
repo: A str holds the repo to load from.
Returns:
A pandas dataframe holds pull request level signals.
"""
pr_level_signals = pd.read_csv(
'../data/%s_pull_requests_signals.csv' % repo)
pr_level_signals['check run passed'] = pr_level_signals[
'check run results'].apply(self._count_check_run_passed)
pr_level_signals['check run failed'] = pr_level_signals[
'check run results'].apply(self._count_check_run_failed)
return pr_level_signals
@staticmethod
def _get_file_level_signals(repo):
"""
Load the file level signals for input repo.
Args:
repo: A str holds the repo to load from.
Returns:
A pandas dataframe holds the file level signals of input repo
with null rows removed.
"""
file_level_signals = pd.read_csv(
'../data/%s_file_level_signals.csv' % repo)
return file_level_signals[file_level_signals['file name'].notna()]
@staticmethod
def _get_file_level_signals_dict(dates, repo):
"""
Load the file level features given repo name and the date range.
Args:
dates: A list of dates.
repo: A str holds the repo name
Returns:
A dict holds the file level features of given repo. Keys are the
dates and values are the dataframes.
"""
file_level_signals_dict = defaultdict(pd.DataFrame)
for date in dates:
date_str = date.strftime(format="%Y_%m_%d")
file_name = '../data/%s_%s_features.csv' % (repo, date_str)
file_level_signals_dict[date_str] = pd.read_csv(file_name)
return file_level_signals_dict
@staticmethod
def get_dates(file_level_signals):
"""
Compute the date range of given file level signals.
Args:
file_level_signals: A dataframe holds the file level signals.
Returns:
A list of dates.
"""
min_date = file_level_signals['pull request closed time'].min()
max_date = file_level_signals['pull request closed time'].max()
start_date = datetime.fromisoformat(min_date[:-1]) \
+ timedelta(days=1)
end_date = datetime.fromisoformat(max_date[:-1])
dates = pd.date_range(start=start_date.strftime("%Y-%m-%d"),
end=end_date.strftime("%Y-%m-%d")) \
.to_pydatetime().tolist()
return dates
@staticmethod
def _get_file_names(files_changes):
"""
Get the file names from the files changes list.
Args:
files_changes: A list of file changes.
Returns:
A list of file names.
"""
file_names = set()
for t in eval(files_changes):
file_name, _, _, _ = t
file_names.add(file_name)
return file_names
@staticmethod
def _get_num_reverted_file(file_names, file_level_signals):
"""
Get the num of files that are involved in CL rollbacks.
Args:
file_names: A list of file names
file_level_signals: A dataframe of file level signals
Returns:
A integer of the num of files that are involved in CL rollbacks.
"""
num_reverted_file = 0
for file_name in file_names:
selected_df = file_level_signals[
file_level_signals['file name'] == file_name]
if selected_df.empty:
continue
if selected_df['reverted pull request id count'].values[0] > 0:
num_reverted_file += 1
return num_reverted_file
@staticmethod
def _get_file_data(pr_id, file_names, file_level_signals, cl_data_dict):
"""
Fill in the file level features of the cl_data_dict and compute the
number of old files
Args:
pr_id: A integer of pull request id.
file_names: A list of file names.
file_level_signals: A dataframe holds the file level signals.
cl_data_dict: A dict of cl_data to fill in.
Returns:
An integer indicating the number of old files.
"""
num_old_files = 0
for i in range(len(file_level_signals)):
file_signals = file_level_signals.iloc[i]
file_name = file_signals['file name']
if file_name in file_names:
file_data = []
for feature in COMMON_FILE_LEVEL_FEATURES:
file_data.append(file_signals[feature])
reverted_cl_rate = \
file_signals['reverted pull request id count'] / \
file_signals['pull request id count']
file_data.append(reverted_cl_rate)
cl_data_dict[pr_id].file_level_features[file_name] = file_data
num_old_files += 1
return num_old_files
@staticmethod
def _get_pr_data(pr_signals, num_files, num_new_files, num_reverted_file):
"""
Get the pull request data.
Args:
pr_signals: A panda series of signals of one pull request.
num_files: An integer of number of files in pull request.
num_new_files: An integer of number of new files in pull request.
num_reverted_file: An integer of number of files that have been
involved in CL rollbacks before.
Returns:
A list of datum of one pull request.
"""
pr_data = []
for feature in COMMON_PR_LEVEL_FEATURES:
pr_data.append(pr_signals[feature])
pr_data.append(num_new_files)
pr_data.append(num_files)
pr_data.append(num_reverted_file)
if num_reverted_file:
pr_data.append(1)
else:
pr_data.append(0)
pr_data.append(num_reverted_file / num_files)
return pr_data
def _get_cl_data_dict(self, pr_level_signals, repo):
"""
Compute the CL data dict.
Args:
pr_level_signals: A dataframe of pull request level signals.
repo: A str of the repo name.
Returns:
A dict holds the CL data. The keys are the CL ids and the values
are one CLData object.
"""
cl_data_dict = defaultdict(CLData)
for index in range(len(pr_level_signals)):
pr_signals = pr_level_signals.iloc[index]
pr_id = pr_signals['pull request id']
reverted_pr_id = pr_signals['reverted pull request id']
if reverted_pr_id != 0:
cl_data_dict[reverted_pr_id].reverted = True
closed_date = datetime.fromisoformat(
pr_signals['pull request closed time'][:-1])\
.strftime(format="%Y_%m_%d")
files_changes = pr_signals['files changes']
file_names = self._get_file_names(files_changes)
num_files = len(file_names)
if not num_files:
continue
file_name = '../data/%s_%s_features.csv' % (repo, closed_date)
if not os.path.exists(file_name):
continue
file_level_signals = pd.read_csv(file_name)
num_reverted_file = self._get_num_reverted_file(file_names,
file_level_signals)
num_old_files = self._get_file_data(pr_id, file_names,
file_level_signals,
cl_data_dict)
num_new_files = num_files - num_old_files
pr_data = self._get_pr_data(
pr_signals, num_files, num_new_files, num_reverted_file)
cl_data_dict[pr_id].pr_level_features = deepcopy(pr_data)
return cl_data_dict
def load_data(self):
"""
Load data from all repos.
Returns:
A dict holds the CL data of all repos. The keys are the repo names
and the values are the CL data.
"""
training_data_dict = defaultdict(list)
for repo in self.repos:
print("Adding %s" % repo)
pr_level_signals = self._get_pr_level_signals(repo)
cl_data_dict = self._get_cl_data_dict(pr_level_signals, repo)
for pr_id in cl_data_dict:
cl_data = cl_data_dict[pr_id]
pr_features = cl_data.pr_level_features
if not pr_features:
continue
file_features = list(cl_data.file_level_features.values())
reverted = cl_data.reverted
training_data_dict[repo].append(
[pr_features, file_features, reverted])
return training_data_dict
def save_data_to_txt(self, training_data_dict):
"""
Save the data of all repos to local txt files.
Args:
training_data_dict: A dict holds the CL data of all repos. The keys are the repo names
and the values are the CL data.
Returns:
None
"""
for repo in training_data_dict:
repo_data = training_data_dict[repo]
with open('../data/%s_data.txt' % repo, 'w') as file:
file.write(str(self.pr_columns))
file.write('\n')
file.write(str(self.file_columns))
file.write('\n')
for datum in repo_data:
file.write(str(datum))
file.write('\n')
def load_data_from_txt(self):
"""
Load the data of all repos from the local txt files.
Returns:
load_pr_columns: A list of pull request level feature names.
load_file_columns: A list of file level feature names.
load_data_dict: A dict holds the CL data of all repos.
The keys are the repo names and the values are the CL data.
"""
load_data_dict = {}
for repo in self.repos:
with open('../data/%s_data.txt' % repo, 'r') as file:
load_pr_columns = eval(file.readline())
load_file_columns = eval(file.readline())
lsts = []
for line in file:
lst = eval(line)
lsts.append(lst)
load_data_dict[repo] = lsts
return load_pr_columns, load_file_columns, load_data_dict
def main():
"""
This main function initializes a DataLoader and load the data from local
features csv files and save the data of all repos to local txt files.
"""
data_loader = DataLoader(REPOS)
training_data_dict = data_loader.load_data()
data_loader.save_data_to_txt(training_data_dict)
if __name__ == "__main__":
main()
| [
"pelny@google.com"
] | pelny@google.com |
4e3c0ef1f25cdcd986f146665468ac1c76395c52 | fac16ad71ac9b09afc9abf0528a98171ac02afc4 | /payment/payments/migrations/0003_category_product.py | ada7734a9c41e562f17f56d3edb03d1a44dd48c7 | [] | no_license | evansmwendwa/payment_gateway | 96dbaf3728ebe4e0875152c96ecfbe7b7004dd98 | afdeab38524ded46d1e557bab696afca9c387e7b | refs/heads/master | 2020-03-10T09:38:25.395169 | 2018-04-12T23:52:34 | 2018-04-12T23:52:34 | 129,314,383 | 0 | 0 | null | 2018-04-12T21:44:34 | 2018-04-12T21:44:33 | null | UTF-8 | Python | false | false | 1,097 | py | # Generated by Django 2.0.3 on 2018-03-26 03:13
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('payments', '0002_auto_20180326_0248'),
]
operations = [
migrations.CreateModel(
name='Category',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('categoryName', models.CharField(max_length=100)),
],
),
migrations.CreateModel(
name='Product',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('productName', models.CharField(max_length=100)),
('productPrice', models.IntegerField()),
('productBrand', models.CharField(max_length=100)),
('productCategory', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='payments.Category')),
],
),
]
| [
"you@example.com"
] | you@example.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.