blob_id stringlengths 40 40 | directory_id stringlengths 40 40 | path stringlengths 3 281 | content_id stringlengths 40 40 | detected_licenses listlengths 0 57 | license_type stringclasses 2 values | repo_name stringlengths 6 116 | snapshot_id stringlengths 40 40 | revision_id stringlengths 40 40 | branch_name stringclasses 313 values | visit_date timestamp[us] | revision_date timestamp[us] | committer_date timestamp[us] | github_id int64 18.2k 668M ⌀ | star_events_count int64 0 102k | fork_events_count int64 0 38.2k | gha_license_id stringclasses 17 values | gha_event_created_at timestamp[us] | gha_created_at timestamp[us] | gha_language stringclasses 107 values | src_encoding stringclasses 20 values | language stringclasses 1 value | is_vendor bool 2 classes | is_generated bool 2 classes | length_bytes int64 4 6.02M | extension stringclasses 78 values | content stringlengths 2 6.02M | authors listlengths 1 1 | author stringlengths 0 175 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
7bf65e7392da61782875463474e98ff06ebb0f75 | 15d049c1d3c427a4e5324f4b1d085a150f955f1a | /app_users/views.py | 022fe473ba9ff5d950afd5efa3900135393040c8 | [] | no_license | isaacnewtonfx/contacts-cicd | 2c597393667e1b678237b39713c538431fa17db1 | af94767ec3ae9e54b131a0932ba00beaa0ac5b8b | refs/heads/master | 2022-05-07T20:43:45.814974 | 2020-01-25T23:09:33 | 2020-01-25T23:09:33 | 233,688,058 | 1 | 0 | null | 2022-04-22T22:57:42 | 2020-01-13T20:34:44 | JavaScript | UTF-8 | Python | false | false | 8,117 | py | import datetime,os,shutil
import pdb,pprint
from django.http import HttpResponse,HttpResponseRedirect,Http404
from django.contrib.auth.decorators import login_required,user_passes_test
from django.contrib.auth import authenticate, login,logout
from django.views.decorators.csrf import csrf_protect
from django.shortcuts import render,get_object_or_404
from django.views.generic import View
from . forms import LoginForm,RegistrationForm,ChangePasswordForm,PersonalDetailsForm,User # the dot means "this current directory"
from django.conf import settings
from django.contrib.auth.models import Group
from django.utils.decorators import method_decorator
from django.contrib import messages
# # test admin @user_passes_test(is_admin)
# def is_admin(user):
# return user.groups.filter(name='admin').exists()
# def is_logged_in(user):
# return user.is_authenticated
# # -------This decorator will apply to all views in this class------
# @method_decorator(user_passes_test(is_logged_in))
# def dispatch(self, *args, **kwargs):
# return super(ManageView, self).dispatch(*args, **kwargs)
# # ------------------------------------------------------------------
class LoginView(View):
def get(self,request):
return render(request, 'app_users/login.html')
def post(self,request):
#pdb.set_trace()
# create a form instance and populate it with data from the request:
form = LoginForm(request.POST,request = request)
if form.is_valid():
# process the data in form.cleaned_data as required
# ...
# redirect to a new URL:
# get username and password
username,password = form.cleaned_data['username'],form.cleaned_data['password']
# authenticate user
user = authenticate(username=username, password=password)
if user is not None:
if user.is_active:
login(request,user)
return HttpResponseRedirect('/contacts')
else:
messages.add_message(request, messages.ERROR, "Sorry, this user is inactive")
return render(request, 'app_users/public_login.html', {'loginform': form})
else:
messages.add_message(request, messages.ERROR, "invalid credentials or inactive user")
return render(request, 'app_users/login.html', {'loginform': form})
return HttpResponseRedirect('/contacts')
else:
return render(request, 'app_users/login.html', {'loginform': form})
class LogoutView(View):
def get(self,request):
logout(request)
messages.add_message(request, messages.SUCCESS, "You are now logged out")
return HttpResponseRedirect('/contacts/users/login')
class RegistrationView(View):
def get(self,request):
#we need to pass an empty form because of the captcha field
form = RegistrationForm()
return render(request, 'app_users/register.html', {'emptyForm':form})
def post(self,request):
# create a form instance and bind it with the request data:
form = RegistrationForm(request.POST,request.FILES)
if form.is_valid():
# creating new user object
user = User.objects.create_user(
username = form.cleaned_data['username'],
email = form.cleaned_data['email'],
password = form.cleaned_data['password'],
)
# update user data
user.first_name = form.cleaned_data['firstname']
user.last_name = form.cleaned_data['lastname']
user.last_login = datetime.datetime.now()
user.date_joined = datetime.datetime.now()
user.is_superuser = False
user.is_staff = True
user.is_active = True
# save the user
user.save()
# add user to the public group if only it exists
if Group.objects.filter(name='public').exists():
g = Group.objects.filter(name='public')
g.user_set.add(user)
# extended profile attributes
user_profile = user.profile
user_profile.middlename = form.cleaned_data['middlename']
user_profile.phone = form.cleaned_data['phone']
user_profile.photo = request.FILES['photo'] if 'photo' in request.FILES else None
user_profile.save()
messages.add_message(request, messages.SUCCESS, "Registration Successful. Your account has now been created")
return HttpResponseRedirect('/contacts')
else:
return render(request, 'app_users/register.html', {'regform': form})
class ManageView(View):
def get(self,request):
return render(request, 'app_users/manage.html')
class ChangeUserPasswordView(View):
def get(self,request):
data = {'show_change_password':True}
return render(request, 'app_users/manage.html', data)
def post(self,request):
# create a form instance and bind it with the request data:
form = ChangePasswordForm(request.POST,request = request)
if form.is_valid():
new_password = form.cleaned_data['new_password']
user = request.user
username = user.username
user.set_password(new_password)
user.save()
# NB! set_password will logout user when it is successful
# hence, Re-authenticate the user with new_password and log them in
user = authenticate(username=username, password=new_password)
login(request, user)
# return to change password page with success message
messages.add_message(request, messages.SUCCESS, "Password was changed Successfully")
return HttpResponseRedirect('/contacts/users/manage/')
else:
return render(request, 'app_users/manage.html', {'show_change_password':True,'change_password_form':form})
class ManagePersonalDetailsView(View):
def get(self,request):
user = request.user
user_profile = user.profile
initial_data = {
'firstname' : user.first_name,
'middlename' : user_profile.middlename,
'lastname' : user.last_name,
'email' : user.email,
'phone' : user_profile.phone
}
#we need to pass an empty form because of the captcha field
form = PersonalDetailsForm(initial = initial_data)
data = {'form':form,'show_personal_details':True}
return render(request, 'app_users/manage.html', data)
def post(self,request):
# create a form instance and bind it with the request data:
form = PersonalDetailsForm(request.POST,request.FILES,request = request)
if form.is_valid():
# get the user object
user = request.user
user.first_name = form.cleaned_data['firstname']
user.last_name = form.cleaned_data['lastname']
user.email = form.cleaned_data['email']
# get the existing user profile data
user_profile = user.profile
old_photo_filename = user_profile.photo.name
user_profile.middlename = form.cleaned_data['middlename']
user_profile.phone = form.cleaned_data['phone']
if 'photo' in request.FILES:
user_profile.photo = request.FILES['photo']
user_profile.hasPhotoUpload = True
else:
user_profile.hasPhotoUpload = False
# Save all changes
user.save()
user_profile.save()
messages.add_message(request, messages.SUCCESS, 'Updated Successfully')
return HttpResponseRedirect('/contacts/users/manage/')
else:
return render(request, 'app_users/manage.html', {'form': form,'show_personal_details':True}) | [
"isaacnewtonfx@gmail.com"
] | isaacnewtonfx@gmail.com |
879f74a6b1320396aec7eeac890e80d9bc6010d2 | f474d500b7da4f4069e24fddcde97783a4f3664b | /vagrantEnv/lib/python3.5/encodings/shift_jis_2004.py | 0e58f25e4692aec46d75fe5c0684e974b5d6ebb5 | [
"Apache-2.0"
] | permissive | Thanh-Lai/chat-bot | 220a0fd6383181f0cdaf732b5c02f645bd960a28 | e3007fa6e034d3cccff4615a7eccf0e75bbc1708 | refs/heads/master | 2020-04-23T09:39:04.509356 | 2019-02-18T04:56:25 | 2019-02-18T04:56:25 | 171,075,880 | 0 | 0 | Apache-2.0 | 2019-02-18T04:56:26 | 2019-02-17T03:00:39 | Python | UTF-8 | Python | false | false | 46 | py | /usr/lib/python3.5/encodings/shift_jis_2004.py | [
"tlai1122@gmail.com"
] | tlai1122@gmail.com |
94b3299de331c59da66c1a70f41ba734bbafbf4c | 83e96abe109d65a922679df75cf3108d9130f88b | /python_20.py | d83b399b76a025b6545e1925b780a1c6685d7d3b | [] | no_license | AStudentCoder/ICS3Uredo | 6c2934154ae4d86f62916e21d4a0987283809709 | dbfd5348fdfb929a6c119aab8c87a801bf4c87b4 | refs/heads/master | 2020-01-23T21:42:47.994742 | 2016-11-24T15:46:59 | 2016-11-24T15:46:59 | 74,686,147 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,619 | py | '''Represent a small bilingual lexicon as a Python dictionary in the following fashion {"merry":"god", "christmas":"jul", "and":"och", "happy":"gott", "new":"nytt", "year":"år"} and use it to translate your Christmas cards from English into Swedish. That is, write a function translate() that takes a list of English words and returns a list of Swedish words.'''
#Tells user to input the saying 'merry christmas and happy new year'
card = input("This is a simple English-Swedish translator for christmas cards. Please type in the phrase 'Merry christmas and happy new year'. Punctuation at the end may be added: ").split()
#lowers all items in card to lowercase
card = [item.lower() for item in card]
#this is the dictionary which translates words from english to swedish
lexicon = {"merry":"God", "christmas":"jul", "and":"och", "happy":"gott", "new":"nytt", "year":"år", "year.":"år.", "year!":"år!"}
#function is defined
def translate(card):
#this empty list will be used later on, to print the translation
empty_list = []
#this goes through each item in card
for item in card:
#if the item is in the lexicon
if item in lexicon:
#add the lexicons value to the empty list
empty_list.append(lexicon[item])
else:
#otherwise just add the item
empty_list.append(item)
#prints that some of the words are not in the lexicon
print ("Some/All of those words were not in our English-Swedish translator!")
#combines the empty list into a string
print (" ".join(empty_list))
#calls the string
translate(card)
| [
"noreply@github.com"
] | noreply@github.com |
0d1cd79ce9a39c39200082549e43260515a74acf | 4e1614393d049b5e07a7fb2204ea47b6ecd341d1 | /pyqt-5 options/scripts/color_picker.py | 653845c090e456278f40908714803e27bc507b68 | [] | no_license | fzerman/pyqt5-options | cee81ae8519a14086f03ce10a906088676b162ba | 47cd649137ab02f951a44c23fdef00ca21e91c99 | refs/heads/main | 2023-04-12T13:47:39.248208 | 2021-05-12T11:15:08 | 2021-05-12T11:15:08 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 729 | py | from PyQt5.QtWidgets import QColorDialog, QPushButton
from PyQt5.QtGui import QColor
class ColorPicker(QPushButton):
def __init__(self,value = [255,255,255,255]):
super().__init__('Open color dialog')
self.rgb = value
self.setToolTip('Opens color dialog')
self.move(10,10)
self.clicked.connect(self.openColorDialog)
def openColorDialog(self):
self.color = QColorDialog.getColor()
if not self.color.isValid():
self.color = False
else:
self.rgb = self.color.getRgb()
self.setStyleSheet("background-color: rgba({},{},{},{})".format( self.rgb[0], self.rgb[1], self.rgb[2], self.rgb[3]))
| [
"noreply@github.com"
] | noreply@github.com |
eeb2122248c6cdc5753e2161399497081b0a4a1c | c32c0bf9c24138a039203087877d8ecd1da81beb | /lib/clients/metaApi/metatraderDemoAccount_client_test.py | 42b10b01e8b2ec1d1382b8ffc6673ce2ccc62598 | [] | no_license | marceluphd/metaapi-python-sdk | c469d4ecd89e6427d86beb613304e1404a2f5b9d | a90760a52f7ecc06e12dfa0d12fc1bc9e6178af5 | refs/heads/master | 2023-01-06T00:25:56.450951 | 2020-11-08T12:01:23 | 2020-11-08T12:01:23 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,138 | py | import responses
import pytest
from ..httpClient import HttpClient
from .metatraderDemoAccount_client import MetatraderDemoAccountClient
PROVISIONING_API_URL = 'https://mt-provisioning-api-v1.agiliumtrade.agiliumtrade.ai'
http_client = HttpClient()
demo_account_client = MetatraderDemoAccountClient(http_client, 'header.payload.sign')
class TestMetatraderDemoAccountClient:
@responses.activate
@pytest.mark.asyncio
async def test_create_mt4(self):
"""Should create new MetaTrader 4 demo from API."""
expected = {
'login': '12345',
'password': 'qwerty',
'serverName': 'HugosWay-Demo3',
'investorPassword': 'qwerty'
}
account = {
'balance': 10,
'email': 'test@test.com',
'leverage': 15,
'serverName': 'server'
}
with responses.RequestsMock() as rsps:
rsps.add(responses.POST, f'{PROVISIONING_API_URL}/users/current/provisioning-profiles/'
'profileId1/mt4-demo-accounts',
json=expected, status=200)
accounts = await demo_account_client.create_mt4_demo_account('profileId1', account)
assert rsps.calls[0].request.url == f'{PROVISIONING_API_URL}/users/current/provisioning-profiles/' + \
'profileId1/mt4-demo-accounts'
assert rsps.calls[0].request.method == 'POST'
assert rsps.calls[0].request.headers['auth-token'] == 'header.payload.sign'
assert accounts == expected
@pytest.mark.asyncio
async def test_not_create_mt4_demo_with_account_token(self):
"""Should not create MetaTrader 4 demo account via API with account token'."""
account_client = MetatraderDemoAccountClient(http_client, 'token')
try:
await account_client.create_mt4_demo_account('', {})
except Exception as err:
assert err.__str__() == 'You can not invoke create_mt4_demo_account method, because you have ' + \
'connected with account access token. Please use API access token from ' + \
'https://app.metaapi.cloud/token page to invoke this method.'
@responses.activate
@pytest.mark.asyncio
async def test_create_mt5(self):
"""Should create new MetaTrader 4 demo from API."""
expected = {
'login': '12345',
'password': 'qwerty',
'serverName': 'HugosWay-Demo3',
'investorPassword': 'qwerty'
}
account = {
'balance': 10,
'email': 'test@test.com',
'leverage': 15,
'serverName': 'server'
}
with responses.RequestsMock() as rsps:
rsps.add(responses.POST, f'{PROVISIONING_API_URL}/users/current/provisioning-profiles/'
'profileId2/mt5-demo-accounts',
json=expected, status=200)
accounts = await demo_account_client.create_mt5_demo_account('profileId2', account)
assert rsps.calls[0].request.url == f'{PROVISIONING_API_URL}/users/current/provisioning-profiles/' + \
'profileId2/mt5-demo-accounts'
assert rsps.calls[0].request.method == 'POST'
assert rsps.calls[0].request.headers['auth-token'] == 'header.payload.sign'
assert accounts == expected
@pytest.mark.asyncio
async def test_not_create_mt5_demo_with_account_token(self):
"""Should not create MetaTrader 5 demo account via API with account token'."""
account_client = MetatraderDemoAccountClient(http_client, 'token')
try:
await account_client.create_mt5_demo_account('', {})
except Exception as err:
assert err.__str__() == 'You can not invoke create_mt5_demo_account method, because you have ' + \
'connected with account access token. Please use API access token from ' + \
'https://app.metaapi.cloud/token page to invoke this method.'
| [
"roman.k.as@agiliumlabs.com"
] | roman.k.as@agiliumlabs.com |
de48afc07698b7b700b7091b68a3e1aee205870d | 9de245acdb1a5a102e6e70d850817aaee6ef861b | /exercises/zAlgorithm.py | c5403887f406639ea1e968f5f769f0b5be392499 | [
"Apache-2.0"
] | permissive | narendraparigi1987/python | 09aac262fc5f9285b105a2bef4a4648a2a1c0d1d | b3e235e7653583c0a1149b75e1527d44c0834de7 | refs/heads/master | 2021-04-03T01:08:13.567406 | 2018-03-09T23:43:12 | 2018-03-09T23:43:12 | 124,603,301 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 933 | py | #!/usr/bin/py
# Head ends here
def zAlgorithm(s):
z = {}
l = 0
r = 0
n = len(s)
for k in range(1,n):
if k > r:
l = r = k
while r < n and s[r] == s[r-l]:
r+=1
z[k] = r -l
r-=1
else:
k1 = k-l
if z[k1] < r-k+1:
z[k] = z[k1]
else:
while r < n and s[r] == s[r-l]:
r+=1
z[k] = r-l
r-=1
k+=1
return z
def stringSimilarity(a):
answer = 0
for i in range(len(a)):
pat = a[i:]
text = pat+'$'+a
if pat[0] == a[0]:
z = zAlgorithm(text)
answer = answer + z[len(pat)+1]
return answer
# Tail starts here
if __name__ == '__main__':
t = int(input())
for i in range(0,t):
a=input()
print(stringSimilarity(a))
#Input
#2
#ababaa
#aa
| [
"naren.parigi@gmail.com"
] | naren.parigi@gmail.com |
51283136a72108723224d37013bae216201001d6 | 1847d6dadc256623b2b838c2013245bb76aa2f66 | /sds/sds/wsgi.py | 200b5143612745741637fb6699d896a708f57113 | [] | no_license | tuckj3569/sds | 9e244c9d12f56015e43b66e0f49a3fe185b87cbd | e37521019f720bed55148dd174d564b64cc7a461 | refs/heads/master | 2022-05-26T10:54:11.006060 | 2018-11-20T00:36:31 | 2018-11-20T00:36:31 | 157,311,329 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 383 | py | """
WSGI config for sds project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/2.0/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "sds.settings")
application = get_wsgi_application()
| [
"38771207+tuckj3569@users.noreply.github.com"
] | 38771207+tuckj3569@users.noreply.github.com |
01b799d7efe96a6343a26b4a2232186869dfae4d | ea00ed279283b2635529ecc03a10b755bb46b525 | /GiveAwayApp/scripts.py | 7c68bd000638cd1843eab2a3fd0c50c82691b7bb | [] | no_license | soboniek/give-away-app | 327772c635aea759798ec43b4b70228a07afe6b1 | eefa16cf28cb509bf5caf9d745ac7ca6edd373f9 | refs/heads/main | 2023-01-24T00:16:28.771216 | 2020-11-23T14:28:17 | 2020-11-23T14:28:17 | 307,805,685 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,955 | py | import datetime
from random import randint
from django.contrib.auth.models import User
from GiveAwayApp.models import Category, Donation, Institution
def create_users():
User.objects.create_user('aaa', email='aaa@aaa.pl', password='aaa')
User.objects.create_user('bbb', email='bbb@bbb.pl', password='bbb')
User.objects.create_user('ccc', email='ccc@ccc.pl', password='ccc')
def create_categories():
for i in range(5):
Category.objects.create(name=f"Name {i}")
def create_donations():
for i in range(5):
Donation.objects.create(quantity=randint(1, 10),
institution=Institution.objects.get(pk=randint(0, 4)),
address=f'Address {i}',
phone_number=randint(100000000, 999999999),
city=f'City {i}',
zip_code=randint(10000, 99999),
pick_up_date=datetime.datetime.today() + datetime.timedelta(days=randint(1, 20)),
pick_up_time=f'{randint(1, 23)}:{randint(1, 59)}',
pick_up_comment=f'Comment {i}',
user=User.objects.all()[0])
def create_institutions():
for i in range(5):
Institution.objects.create(name=f'Name {i}',
description=f'Description {i}',
type=randint(1, 3))
def create_donation_categories():
for i in range(1, 6):
donation = Donation.objects.get(id=i)
for j in range(3):
cat = Category.objects.get(id=randint(1, 5))
donation.categories.add(cat)
def create_institution_categories():
for i in range(1, 6):
institution = Institution.objects.get(id=i)
for j in range(3):
cat = Category.objects.get(id=randint(1, 5))
institution.categories.add(cat)
| [
"olga.lepecka@gmail.com"
] | olga.lepecka@gmail.com |
2d47233cae6785171b2a36fbf2d434ff58c079fb | 5f95ea2d1f2d847121b94631c4e92266de507fb1 | /src/modeling.py | 0e8297662b522287f0f2fa05a3bc777a33cd5549 | [
"MIT"
] | permissive | katsu1110/numeraizer | bedcbfcfa5ad1ca24fb7f2bd183bccd534862bfa | 49f788a7e53605263351516171ba4d6d6e7cb8df | refs/heads/main | 2023-08-07T12:50:04.851786 | 2021-09-16T14:32:06 | 2021-09-16T14:32:06 | 291,437,811 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 10,593 | py | # libraries
import numpy as np # linear algebra
import pandas as pd # data processing, CSV file I/O (e.g. pd.read_csv)
import gc
from tqdm.auto import tqdm
import joblib
import pathlib
import glob
import json
import scipy
from scipy.stats import skew, kurtosis
from sklearn import decomposition, preprocessing, linear_model, svm
from multiprocessing import Pool, cpu_count
import time
import requests as re
import datetime
from dateutil.relativedelta import relativedelta, FR
import operator
import xgboost as xgb
import lightgbm as lgb
# visualize
import matplotlib.pyplot as plt
import matplotlib.style as style
from matplotlib_venn import venn2, venn3
import seaborn as sns
from matplotlib import pyplot
from matplotlib.ticker import ScalarFormatter
sns.set_context("talk")
style.use('seaborn-colorblind')
import warnings
warnings.simplefilter('ignore')
"""Utils
"""
def make_group(feature_df, key='friday_date'):
"""group maker for rank learning
"""
return feature_df.groupby(key).size().to_frame('size')['size'].to_numpy()
def make_dataset(feature_df, features, target='target20_d', SUFFIX='val', ):
"""Make dataset
"""
# ---------------------------------------------------------
# making dataset
# ---------------------------------------------------------
# train set
if SUFFIX == 'val': # for validation
train_set = {
'X': feature_df.query('data_type == "train"')[features],
'y': feature_df.query('data_type == "train"')[target].astype(np.float64),
'g': make_group(feature_df.query('data_type == "train"'))
}
else: # full
data_types = ['train', 'validation']
train_set = {
'X': feature_df.query('data_type in @data_types')[features],
'y': feature_df.query('data_type in @data_types')[target].astype(np.float64),
'g': make_group(feature_df.query('data_type in @data_types'))
}
assert train_set['y'].isna().sum() == 0
# valid set
val_df_ = feature_df.query('data_type == "validation"').dropna(subset=[target]).copy()
val_set = {
'X': val_df_[features],
'y': val_df_[target].astype(np.float64),
'g': make_group(val_df_)
}
assert train_set['y'].isna().sum() == 0
assert val_set['y'].isna().sum() == 0
# test set
test_set = {
'X': feature_df.query('data_type == "live"')[features],
'g': make_group(feature_df.query('data_type == "live"'))
}
return train_set, val_set, test_set
def hypara_dispatcher(MODEL='LGB'):
"""Dispatch hyperparameter
"""
# parameters
if MODEL == 'LGB':
params = {
'n_estimators': 10000,
'objective': 'regression',
'boosting_type': 'gbdt',
'max_depth': 7,
'learning_rate': 0.01,
'subsample': 0.72,
'subsample_freq': 4,
'feature_fraction': 0.1,
'lambda_l1': 1,
'lambda_l2': 1,
'seed': 46,
'verbose': -1,
# 'device': 'gpu'
}
params["metric"] = "rmse"
elif MODEL == 'XGB':
params = {
'colsample_bytree': 0.1,
'learning_rate': 0.1,
'max_depth': 4,
'subsample': 1,
'min_child_weight': 4,
'gamma': 0.24,
'alpha': 1,
'lambda': 1,
'seed': 46,
'n_estimators': 10000,
'tree_method': 'gpu_hist' # Let's use GPU for a faster experiment
}
# params["objective"] = 'rank:pairwise'
elif MODEL == 'MLP':
params = {
'input_dropout': 0.0,
'hidden_layers': 3,
'hidden_units': 256,
'embedding_out_dim': 4,
'hidden_activation': 'relu',
'hidden_dropout': 0.01,
'gauss_noise': 0.01,
'norm_type': 'layer', # layer
'optimizer': {'type': 'adam', 'lr': 1e-3},
'batch_size': 1024,
'epochs': 100
}
elif MODEL == 'ridge':
params = {
'alpha': 100
, 'fit_intercept': True
, 'max_iter': 10000
, 'random_state': 46
}
elif MODEL == 'beyesianridge':
params = {
'n_iter': 10000
}
elif MODEL == 'lasso':
params = {
'alpha': 0.001
, 'fit_intercept': True
, 'max_iter': 10000
, 'random_state': 46
}
elif MODEL == 'svm':
params = {
'C': 100
}
return params
def model_trainer(train_set, val_set, MODEL='LGB', SUFFIX='val', SAVE=False):
"""Train model
"""
logger.info(f'Training {MODEL} in {SUFFIX} mode...')
# get hyperparameters
params = hypara_dispatcher(MODEL)
# fit
if MODEL == 'LGB':
# train data
dtrain_set = lgb.Dataset(
train_set['X'].values
, train_set['y'].values
, feature_name=features
)
if SUFFIX == 'val':
# val data
dval_set = lgb.Dataset(
val_set['X'].values
, val_set['y'].values
, feature_name=features
)
# train
model = lgb.train(
params
, dtrain_set
, valid_sets=[dtrain_set, dval_set]
, early_stopping_rounds=100
, verbose_eval=100
)
else:
# train
model = lgb.train(
params
, dtrain_set
)
elif MODEL == 'XGB':
# model
model = xgb.XGBRegressor(**params)
if SUFFIX == 'val':
# train
model.fit(
train_set['X'], train_set['y'],
eval_set=[(val_set['X'], val_set['y'])],
verbose=500,
early_stopping_rounds=100,
)
else:
# train
model.fit(
train_set['X'], train_set['y'],
verbose=500,
)
elif MODEL == 'XGBRank':
# model
model = xgb.XGBRanker(**params)
if SUFFIX == 'val':
model.fit(
train_set['X'], train_set['y'],
eval_set=[(val_set['X'], val_set['y'])],
group=train_set['g'],
eval_group=[val_set['g']],
verbose=100,
early_stopping_rounds=100,
)
else:
model.fit(
train_set['X'], train_set['y'],
group=train_set['g'],
verbose=100,
)
# save model
if SAVE:
if MODEL[-1] == 'B':
# save via joblib
joblib.dump(model, f'{OUTPUT_DIR}/{target}_{MODEL}_model_{SUFFIX}.pkl')
logger.info(f'{MODEL} {SUFFIX}_model for {target} saved!')
return model
def get_feature_importance(model, features, MODEL='LGB'):
"""Get feature importance
"""
# feature importance
fi_df = pd.DataFrame()
fi_df['features'] = features
fi_df['importance'] = np.nan
# LGB
if MODEL == 'LGB':
fi_df['importance'] = model.feature_importance(importance_type="gain")
# XGB
elif 'XGB' in MODEL:
importance = model.get_booster().get_score(importance_type='gain')
importance = sorted(importance.items(), key=operator.itemgetter(1))
fi_df = pd.DataFrame(importance, columns=['features', 'importance'])
return fi_df
def fit_model(feature_df, features, targets=['target_20d', 'target_4d'], MODEL='LGB', ):
"""Fit model
"""
# fit
fi_df = pd.DataFrame()
fi_df['features'] = features
for target in tqdm(targets):
logger.info(' ======================== ')
logger.info(f'{MODEL}: predicting {target}...!')
logger.info(' ======================== ')
# make datasets
train_set, val_set, test_set = make_dataset(feature_df, features, target, SUFFIX='val')
assert train_set['y'].isna().sum() == 0
assert val_set['y'].isna().sum() == 0
# train with validation data
model = model_trainer(train_set, val_set, MODEL, SUFFIX='val', SAVE=True)
# feature importance
fi_df_ = get_feature_importance(model, features, MODEL)
fi_df = fi_df.merge(
fi_df_.rename(columns={'importance': f'{target}_{MODEL}'})
, how='left'
, on='features'
)
# val model prediction
sub_df[f'{target}_{MODEL}'] = model.predict(sub_df[features])
# full model training
train_set = {
'X': feature_df.loc[feature_df['data_type'].isin(['train', 'validation']), features],
'y': feature_df.loc[feature_df['data_type'].isin(['train', 'validation']), target].astype(np.float32)
}
params_full = params.copy()
params_full['n_estimators'] = 201
dtrain_set = lgb.Dataset(
train_set['X'].values, train_set['y'].values
, feature_name=features
)
model = lgb.train(
params_full
, dtrain_set
, verbose_eval=500
)
# full model prediction
live_pred_val = sub_df.loc[sub_df['data_type'] == 'live', f'{target}_{MODEL}'].values
live_pred = model.predict(sub_df.loc[sub_df['data_type'] == 'live', features])
sub_df.loc[sub_df['data_type'] == 'live', f'{target}_{MODEL}'] = 0.3*live_pred_val + 0.7*live_pred
# save
joblib.dump(model, f'{OUTPUT_DIR}/{target}_{MODEL}_model_full.pkl')
logger.info(f'{MODEL} full-model for {target} saved!')
def plot_feature_importance(MODEL='LGB', targets=['target_20d', 'target_4d', ], top_n=25):
"""Plot feature importance
"""
fig, ax = plt.subplots(1, len(targets), figsize=(16, 12))
for i, target in tqdm(enumerate(targets)):
pred_col = f'{target}_{MODEL}'
sns.barplot(
x=pred_col
, y='features'
, data=fi_df.sort_values(by=pred_col, ascending=False).iloc[:top_n]
, ax=ax[i]
)
if i > 0:
ax[i].set_ylabel('')
if i == 1:
ax[i].set_xlabel('importance')
else:
ax[i].set_xlabel('')
ax[i].set_title(pred_col)
plt.tight_layout() | [
"katsuhisa@KatsuhisanoMacBook-Pro.local"
] | katsuhisa@KatsuhisanoMacBook-Pro.local |
689cfdb0bc848c8a3eae966b31ec04fd9e24f7f6 | 0bb27c2d2e1f7e44537515ae31f8e1713ece8407 | /5-6. 라이브러리/05-6-06.py | 697e30a7580686b3e8efcaa896a5ab8948a98951 | [] | no_license | calmahn/jump_to_python | f00d5aa11d1e2abe82f3698a403a288ed03acdb8 | 164ea4ef72f9a1ed4bc62f8580404d833ba4e442 | refs/heads/master | 2023-04-14T08:00:00.552291 | 2021-04-27T11:48:03 | 2021-04-27T11:48:03 | 345,359,084 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 101 | py | import tempfile
filename = tempfile.mkstemp()
print(filename)
f = tempfile.TemporaryFile()
f.close() | [
"acy824@naver.com"
] | acy824@naver.com |
5b4c0437db6470f56799c36a888f998b93d784ab | f28bb4392681912679b41b7e2d1994404e68f772 | /blogpy_articles/migrations/0001_initial.py | 06f1ab5a22fd1e720c396dfc7a2916eaa90fb772 | [] | no_license | erfanmorsali/my_blogpy | 7385f3c15f02ff1c4fcfce11792b29b643d7df6f | 15bb670fe866612ee86dbe2968c708263a24a69e | refs/heads/master | 2023-02-22T08:06:24.704781 | 2021-01-30T09:08:45 | 2021-01-30T09:08:45 | 307,731,993 | 0 | 2 | null | 2020-11-03T15:50:10 | 2020-10-27T14:45:59 | Python | UTF-8 | Python | false | false | 1,118 | py | # Generated by Django 3.1.2 on 2020-10-27 19:43
from django.conf import settings
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
initial = True
dependencies = [
migrations.swappable_dependency(settings.AUTH_USER_MODEL),
]
operations = [
migrations.CreateModel(
name='Article',
fields=[
('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')),
('title', models.CharField(max_length=200, verbose_name='عنوان مقاله')),
('description', models.TextField(verbose_name='توضیحات')),
('image', models.ImageField(upload_to='articles/', verbose_name='تصویر مقاله')),
('created_at', models.DateTimeField(auto_now_add=True, verbose_name='تاریخ ثبت مقاله')),
('author', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL, verbose_name='نویسنده')),
],
),
]
| [
"erfanmorsalidev@gmail.com"
] | erfanmorsalidev@gmail.com |
747b7dc02723dca66a8f5383174530e1e272eb83 | 6a6a272013dae1faba8cd31058a7b0bbd6c2b048 | /book_dongbinna_with_python/chapter09최단거리/미래도시.py | 88388602793b32891bcd0768b3ba782130c197ed | [] | no_license | daegu-algo-party-210824/hojin_algo | b21addae1d5d6a827877f99c20f9498bd328e877 | 1841f0ab9be8a9b4407d8cb9f9fc348373e78f8a | refs/heads/main | 2023-08-17T19:19:51.621761 | 2021-10-17T21:08:39 | 2021-10-17T21:08:39 | 399,362,538 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,251 | py | INF = int(1e9)
# 노드의 개수 및 간선의 개수를 입력받기
n, m = map(int, input().split())
# 2차원 리스트를 만들고, 모든 값을 무한으로 초기화
graph = [[INF] * (n + 1) for _ in range(n + 1)]
# 자기 자신에서 자기 자신으로 가는 비용은 0으로 초기화
for a in range(1, n + 1):
for b in range(1, n + 1):
if a == b:
graph[a][b] = 0
# 각 간선에 대한 정보를 입력받아, 그 값으로 초기화
for _ in range(m):
# A와 B가 서로에게 가는 비용은 1이라고 설정
a, b = map(int, input().split())
graph[a][b] = 1
graph[b][a] = 1
# 거쳐 갈 노드 X 와 최종 목적지 노드 K를 입력받기
x, k = map(int, input().split())
# 점화식에 따라 플로이드 워셜 알고리즘을 수행
for k in range(1, n + 1):
for a in range(1, n + 1):
for b in range(1, n + 1):
graph[a][b] = min(graph[a][b], graph[a][k] + graph[k][b])
# 수행된 결과를 출력
distance = graph[1][k] + graph[k][x]
# 도달할 수 없는 경우, -1을 출력
if distance >= INF:
print("-1")
# 도달할 수 있다면, 최단거리를 출력
else:
print(distance)
"""
5 7
1 2
1 3
1 4
2 4
3 4
3 5
4 5
4 5
4 2
1 3
2 4
3 4
""" | [
"hojin9car@gmail.com"
] | hojin9car@gmail.com |
77eaad8fff2a0f72251107e65f2fd3b4fb38914c | 4dc4edd6cb8a6e895cce46ac0c4ed7f9655cc0fd | /Array.py | 616ee4b264052741e83c7d02f25da045ce5e28e5 | [
"MIT"
] | permissive | KaloyanDragiev/Python_Jenkins_Splunk | 98380a68bb5002dd83c568ced18b371e6c1f77c7 | 8d0c6139fc4121baed0f0208f56ad6933a728db8 | refs/heads/main | 2023-04-10T16:49:15.387056 | 2021-04-27T12:44:49 | 2021-04-27T12:44:49 | 360,524,650 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 71 | py | cars = ["Mercedes-Benz", "Audi", "BMW"]
for x in cars:
print(x)
| [
"noreply@github.com"
] | noreply@github.com |
2d5c82c21a0762d88be5c748f3c0444cf149e6ba | e61467cee2ace448cd138cfe3891865d149e18aa | /freelancer/settings.py | 7c1b74b623f180adfcb49a58e06f760c01365f68 | [] | no_license | leonardo1909/freelancer | 159958c035717682a8b9a7487194d54748f32347 | 171615c957c9d9fcf25a2a5a37d0986b06c6608b | refs/heads/master | 2022-12-05T03:53:41.037631 | 2020-08-23T21:50:31 | 2020-08-23T21:50:31 | 289,768,569 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,185 | py | """
Django settings for freelancer project.
Generated by 'django-admin startproject' using Django 3.1.
For more information on this file, see
https://docs.djangoproject.com/en/3.1/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/3.1/ref/settings/
"""
from pathlib import Path
# Build paths inside the project like this: BASE_DIR / 'subdir'.
BASE_DIR = Path(__file__).resolve(strict=True).parent.parent
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/3.1/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 'lw$thitde!4mc*a=(g8nul6-6k5bl!n#4*-e9rphq^nt9jzxsg'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'habilidades',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'freelancer.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'freelancer.wsgi.application'
# Database
# https://docs.djangoproject.com/en/3.1/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': BASE_DIR / 'db.sqlite3',
}
}
# Password validation
# https://docs.djangoproject.com/en/3.1/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/3.1/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/3.1/howto/static-files/
STATIC_URL = '/static/'
MEDIA_URL = '/media/'
STATIC_ROOT = 'vol/web/static'
MEDIA_ROOT = 'vol/web/media'
| [
"leonardo.jose94@outlook.com"
] | leonardo.jose94@outlook.com |
33c9ea501c3e9f897d843332978df996557de4fc | 905a2ac66149377da7832c7457201af66da4d846 | /order/tests.py | 3f1ae1045eb4c8558a88cb3219794371ee1ad788 | [] | no_license | no0xgold/one-page-ecommerce-web-app | d185ad0ba2369c56e9482471328f1432a90a65e2 | ab56f63e59e68365753b4b5c4f3a0faf95a21a8e | refs/heads/main | 2023-04-17T20:39:20.845733 | 2021-04-30T13:21:35 | 2021-04-30T13:21:35 | 342,781,407 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 623 | py |
from django.conf import settings
from django.contrib.auth import get_user_model
from django.test import TestCase
User=get_user_model()
# Create your tests here.
class OrderTestCase(TestCase):
def setUp(self):
user_a_pass ='amira12234'
self.user_a_pass=user_a_pass
user_a = User(username='amira', email='amira@amira.com')
user_a.is_staff= True
user_a.is_superuser = True
user_a.set_password(user_a_pass)
user_a.save()
self.user_a = user_a
def test_create_order(self):
obj = Order.objects.create(user=self.user_a, product=product_a)
| [
"noxraktor@gmail.com"
] | noxraktor@gmail.com |
7175e51e0ffb40cc2db8f90702cd73f0d831deeb | 415a614a54eae170e926154593bab34aac4695fb | /python-dictionary/small1.py | 9b530d8d8c1ce6e84a971309d11e75a936d06b27 | [] | no_license | DasomAnH/python | 434f8d16ad5a437c80efc62c0ae16ea6437a299b | 14b248f71bc0784f433e5e8e38703f97c5d98cb5 | refs/heads/master | 2022-12-12T22:35:56.726214 | 2020-08-31T16:32:13 | 2020-08-31T16:32:13 | 289,390,734 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 435 | py | phonebook_dict = {
'Alice': '703-493-1834',
'Bob': '857-384-1234',
'Elizabeth': '484-584-2923'
}
# retrieve value
person = phonebook_dict['Elizabeth']
print(person)
# add a value
phonebook_dict['kareem'] = '938-489-1234'
# delete
phonebook_dict['Alice'] = 'number hot found'
del phonebook_dict['Alice']
# removed_contact = phonebook_dict.pop('Alice')
# edit
phonebook_dict['Bob'] = "968-345-2345"
print(phonebook_dict)
| [
"ektha116@gmail.com"
] | ektha116@gmail.com |
21905c5fbd4025d75f10237b226ebf37b75e9435 | f4d1492bf4e01dabcac68f2244f9fd540c21ca29 | /Card.py | 127c772f1389660dbb333434d7f7ac87d723f7df | [] | no_license | jordi2105/LabyrinthBot | 9c346e70e9f98f1d5d79509b907bfccf72b3a187 | 57124118b7ae2997b0f157c7fe75fec3caaf875d | refs/heads/master | 2022-12-03T20:55:56.944994 | 2020-09-02T10:56:29 | 2020-09-02T10:56:29 | 264,283,068 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 654 | py | import Objective
import pygame as p
import copy
class Card:
def __init__(self, objective: Objective=None, image_file_url: str=None):
self.objective = objective
if image_file_url is not None:
image = p.image.load(image_file_url)
self.image = p.transform.scale(image, (100, 150))
def copy(self):
copy_obj = Card()
for name, attr in self.__dict__.items():
if hasattr(attr, 'copy') and callable(getattr(attr, 'copy')):
copy_obj.__dict__[name] = attr.copy()
else:
copy_obj.__dict__[name] = copy.deepcopy(attr)
return copy_obj
| [
"jordi.verheul@hotmail.com"
] | jordi.verheul@hotmail.com |
5361f7d7c94132ad1719f0f21b8ce0657580b815 | f8bdc85e59cc703ec3fa771ea2abac9950b9ea83 | /server/pathgenerator.py | 6cc53fa91731d2f4d19e2b1b4c6ff76ba1053b8c | [] | no_license | subash-a/protosketch | 7b31fec37fa2246a5cccb4a17a8febb094b4c7a6 | f4d933e7e609bc6a1296cead8fd0c30e1aa0e22a | refs/heads/master | 2020-03-30T12:35:37.976140 | 2014-07-22T17:24:28 | 2014-07-22T17:24:28 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 67 | py | import uuid
def getNewPrototypeId():
return str(uuid.uuid4())
| [
"s7subash@gmail.com"
] | s7subash@gmail.com |
6a1f1091be991369d46e65e58cffcef7aa13c248 | 2d80f83b1687a890c0016c677e5941472a618ba8 | /hanbitco/__init__.py | 022ef4c5e728a5d8b3cd5e4968ed8af39c9d06ab | [
"MIT"
] | permissive | plutusds/hanbitco-api-python | d1ac616b853093069b1a99d1cf168b018e8c9ac7 | 0783d5f9be0668b42eba295a736cea08191806b3 | refs/heads/main | 2023-04-03T12:48:09.629754 | 2021-03-24T03:22:28 | 2021-03-24T03:22:28 | 348,755,381 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 150 | py | from hanbitco.constants import OrderType, OrderSide, OrderStatus
from hanbitco.utils import create_order_payload
from hanbitco.api import HanbitcoAPI
| [
"kevink1103@gmail.com"
] | kevink1103@gmail.com |
edfb5453073a6d9575cdaf11a8e4117f7ae0ec0d | 5e05c6ec892d9a6bc33c0c0a9b6ce4c7135a83f4 | /cristianoronaldoyopmailcom_299/settings.py | d5a0c910720d8dd82153b4b4433f70e3d17e090e | [] | no_license | payush/cristianoronaldoyopmailcom-299 | 54eb5118840ea7ea68f077ffd7032a62a79880f3 | 52e5bb6ad599605b8cdf1088f9d7cdcf7c1a0265 | refs/heads/master | 2020-03-23T14:23:17.476546 | 2018-07-20T06:30:07 | 2018-07-20T06:30:07 | 141,672,766 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,157 | py | """
Django settings for cristianoronaldoyopmailcom_299 project.
Generated by 'django-admin startproject' using Django 1.11.5.
For more information on this file, see
https://docs.djangoproject.com/en/1.11/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/1.11/ref/settings/
"""
import os
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/1.11/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = 't!!vo0zfzvwkp-_r@$vuqjc=hanbxi^#jl1w9*^z8m(q)mlke8'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = []
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'django.contrib.sites'
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'cristianoronaldoyopmailcom_299.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'cristianoronaldoyopmailcom_299.wsgi.application'
# Database
# https://docs.djangoproject.com/en/1.11/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),
}
}
# Password validation
# https://docs.djangoproject.com/en/1.11/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/1.11/topics/i18n/
LANGUAGE_CODE = 'en-us'
TIME_ZONE = 'UTC'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/1.11/howto/static-files/
STATIC_URL = '/static/'
import environ
env = environ.Env()
ALLOWED_HOSTS = ['*']
SITE_ID = 1
MIDDLEWARE += ['whitenoise.middleware.WhiteNoiseMiddleware']
DATABASES = {
'default': env.db()
}
AUTHENTICATION_BACKENDS = (
'django.contrib.auth.backends.ModelBackend',
'allauth.account.auth_backends.AuthenticationBackend'
)
STATIC_ROOT = os.path.join(BASE_DIR, "staticfiles")
STATICFILES_DIRS = [
os.path.join(BASE_DIR, 'static')
]
STATICFILES_STORAGE = 'whitenoise.storage.CompressedManifestStaticFilesStorage'
LOCAL_APPS = [
'home',
]
THIRD_PARTY_APPS = [
'rest_framework',
'rest_framework.authtoken',
'bootstrap4',
'allauth',
'allauth.account',
'allauth.socialaccount',
'allauth.socialaccount.providers.google',
]
INSTALLED_APPS += LOCAL_APPS + THIRD_PARTY_APPS
# allauth
ACCOUNT_EMAIL_REQUIRED = True
ACCOUNT_AUTHENTICATION_METHOD = 'email'
ACCOUNT_USERNAME_REQUIRED = False
ACCOUNT_EMAIL_VERIFICATION = None
LOGIN_REDIRECT_URL = '/'
| [
"ayushpuroheet@gmail.com"
] | ayushpuroheet@gmail.com |
4334d50846ce9d51cd46cd5202f5b2cccab0c0b3 | 64806194523cf79e3d948f2e3e4bf28d1002f6d3 | /mkb/evaluation/evaluation.py | b6c655ad1573d8f4bb6183232e9ee1bbe7d415a8 | [
"MIT"
] | permissive | EmanuelaBoros/mkb | 6cfef3797db9d19d5e93c0694f6c344dc9a6899d | 3a75f10a5cf86fb74bb349530982130b6e35740a | refs/heads/master | 2023-06-27T12:46:50.041189 | 2021-08-06T17:23:05 | 2021-08-06T17:23:05 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 16,019 | py | from creme import stats
import pandas as pd
from torch.utils import data
import torch
import collections
from ..datasets import base
from ..utils import Bar
__all__ = ['Evaluation']
class Evaluation:
"""Evaluate model on selected dataset.
Evaluate metrics Hits@1, Hits@3, Hits@10, MRR, MR on entities, relations and tails.
Returns distincts metrics for link prediction ie, entities or relations and relation
predictions.
Parameters:
entities (dict): Entities of the dataset.
relations (dict): Relations of the dataset.
batch_size (int): Size of the batch.
true_triples (list): Available triplets to filter_bias metrics. If not specified, `Evaluation`
will mesure raw metrics. Usually we filter_bias triplets based on train, validation and test
datasets.
device (str): cpu or cuda.
num_workers (str): Number of workers for pytorch dataset.
Example:
>>> from mkb import datasets
>>> from mkb import evaluation
>>> from mkb import models
>>> from mkb import losses
>>> from mkb import sampling
>>> import torch
>>> _ = torch.manual_seed(42)
>>> train = [
... (0, 0, 1),
... (0, 1, 1),
... (2, 0, 3),
... (2, 1, 3),
... ]
>>> valid = [
... (0, 0, 1),
... (2, 1, 3),
... ]
>>> test = [
... (0, 0, 1),
... (2, 1, 3),
... ]
>>> entities = {
... 'e0': 0,
... 'e1': 1,
... 'e2': 2,
... 'e3': 3,
... }
>>> relations = {
... 'r0': 0,
... 'r1': 1,
... }
>>> dataset = datasets.Dataset(
... train = train,
... valid = valid,
... test = test,
... entities = entities,
... relations = relations,
... batch_size = 2,
... seed = 42,
... shuffle = False,
... )
>>> negative_sampling = sampling.NegativeSampling(
... size = 2,
... train_triples = dataset.train,
... entities = dataset.entities,
... relations = dataset.relations,
... seed = 42,
... )
>>> model = models.RotatE(hidden_dim=3, entities=dataset.entities,
... relations=dataset.relations, gamma=1)
>>> optimizer = torch.optim.Adam(
... filter(lambda p: p.requires_grad, model.parameters()),
... lr = 0.5,
... )
>>> loss = losses.Adversarial(alpha=0.5)
>>> for _ in range(5):
... for data in dataset:
... sample, weight, mode = data['sample'], data['weight'], data['mode']
... positive_score = model(sample)
... negative_sample = negative_sampling.generate(sample=sample, mode=mode)
... negative_score = model(sample, negative_sample, mode)
... loss(positive_score, negative_score, weight).backward()
... _ = optimizer.step()
>>> model = model.eval()
>>> validation = evaluation.Evaluation(true_triples=train + valid + test,
... entities=entities, relations=relations, batch_size=2)
>>> validation.eval(model=model, dataset=test)
{'MRR': 0.5417, 'MR': 2.25, 'HITS@1': 0.25, 'HITS@3': 1.0, 'HITS@10': 1.0}
>>> validation.eval_relations(model=model, dataset=test)
{'MRR_relations': 1.0, 'MR_relations': 1.0, 'HITS@1_relations': 1.0, 'HITS@3_relations': 1.0, 'HITS@10_relations': 1.0}
>>> validation.detail_eval(model=model, dataset=test, threshold=1.5)
head tail metadata
MRR MR HITS@1 HITS@3 HITS@10 MRR MR HITS@1 HITS@3 HITS@10 frequency
relation
1_1 0.0000 0.0 0.0 0.0 0.0 0.0000 0.0 0.0 0.0 0.0 0.0
1_M 0.0000 0.0 0.0 0.0 0.0 0.0000 0.0 0.0 0.0 0.0 0.0
M_1 0.0000 0.0 0.0 0.0 0.0 0.0000 0.0 0.0 0.0 0.0 0.0
M_M 0.6667 2.0 0.5 1.0 1.0 0.4167 2.5 0.0 1.0 1.0 1.0
>>> validation.types_relations(model = model, dataset=test, threshold=1.5)
{'r0': 'M_M', 'r1': 'M_M'}
References:
1. [RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space](https://github.com/DeepGraphLearning/KnowledgeGraphEmbedding)
"""
def __init__(self, entities, relations, batch_size, true_triples=[], device='cpu', num_workers=1):
self.entities = entities
self.relations = relations
self.true_triples = true_triples
self.batch_size = batch_size
self.device = device
self.num_workers = num_workers
def _get_test_loader(self, triples, mode):
test_dataset = base.TestDataset(
triples=triples, true_triples=self.true_triples, entities=self.entities,
relations=self.relations, mode=mode)
return data.DataLoader(
dataset=test_dataset, batch_size=self.batch_size, num_workers=self.num_workers,
collate_fn=base.TestDataset.collate_fn)
def get_entity_stream(self, dataset):
"""Get stream dedicated to link prediction."""
head_loader = self._get_test_loader(triples=dataset, mode='head-batch')
tail_loader = self._get_test_loader(triples=dataset, mode='tail-batch')
return [head_loader, tail_loader]
def get_relation_stream(self, dataset):
"""Get stream dedicated to relation prediction."""
test_dataset = base.TestDatasetRelation(
triples=dataset, true_triples=self.true_triples, entities=self.entities,
relations=self.relations)
return data.DataLoader(
dataset=test_dataset, batch_size=self.batch_size, num_workers=self.num_workers,
collate_fn=base.TestDatasetRelation.collate_fn)
def eval(self, model, dataset):
"""Evaluate selected model with the metrics: MRR, MR, HITS@1, HITS@3, HITS@10"""
metrics = collections.OrderedDict({
metric: stats.Mean()
for metric in ['MRR', 'MR', 'HITS@1', 'HITS@3', 'HITS@10']
})
with torch.no_grad():
for test_set in self.get_entity_stream(dataset):
metrics = self.compute_score(
model=model,
test_set=test_set,
metrics=metrics,
device=self.device
)
return {name: round(metric.get(), 4) for name, metric in metrics.items()}
def eval_relations(self, model, dataset):
metrics = collections.OrderedDict({
f'{metric}': stats.Mean()
for metric in ['MRR', 'MR', 'HITS@1', 'HITS@3', 'HITS@10']
})
with torch.no_grad():
metrics = self.compute_score(
model=model,
test_set=self.get_relation_stream(dataset),
metrics=metrics,
device=self.device
)
return {f'{name}_relations': round(metric.get(), 4) for name, metric in metrics.items()}
@classmethod
def compute_score(cls, model, test_set, metrics, device):
training = False
if model.training:
model = model.eval()
training = True
bar = Bar(dataset=test_set, update_every=1)
bar.set_description('Evaluation')
for data in bar:
sample = data['sample'].to(device)
negative_sample = data['negative_sample'].to(device)
filter_bias = data['filter_bias'].to(device)
mode = data['mode']
if mode == 'head-batch' or mode == 'tail-batch':
score = model(
sample=sample,
negative_sample=negative_sample,
mode=mode
)
elif mode == 'relation-batch':
score = model(negative_sample)
score += filter_bias
argsort = torch.argsort(score, dim=1, descending=True)
if mode == 'head-batch':
positive_arg = sample[:, 0]
if mode == 'relation-batch':
positive_arg = sample[:, 1]
elif mode == 'tail-batch':
positive_arg = sample[:, 2]
batch_size = sample.size(0)
for i in range(batch_size):
# Notice that argsort is not ranking
ranking = (argsort[i, :] == positive_arg[i]).nonzero()
assert ranking.size(0) == 1
ranking = 1 + ranking.item()
# ranking + 1 is the true ranking used in evaluation metrics
metrics['MRR'].update(1.0/ranking)
metrics['MR'].update(ranking)
metrics['HITS@1'].update(
1.0 if ranking <= 1 else 0.0)
metrics['HITS@3'].update(
1.0 if ranking <= 3 else 0.0)
metrics['HITS@10'].update(
1.0 if ranking <= 10 else 0.0)
if training:
model = model.train()
return metrics
@classmethod
def compute_detailled_score(cls, model, test_set, metrics, types_relations, device):
training = False
if model.training:
model = model.eval()
training = True
bar = Bar(dataset=test_set, update_every=1)
bar.set_description('Evaluation')
for data in bar:
sample = data['sample'].to(device)
negative_sample = data['negative_sample'].to(device)
filter_bias = data['filter_bias'].to(device)
mode = data['mode']
score = model(
sample=sample,
negative_sample=negative_sample,
mode=mode,
)
score += filter_bias
argsort = torch.argsort(score, dim=1, descending=True)
if mode == 'head-batch':
positive_arg = sample[:, 0]
elif mode == 'tail-batch':
positive_arg = sample[:, 2]
batch_size = sample.size(0)
for i in range(batch_size):
# Notice that argsort is not ranking
ranking = (argsort[i, :] == positive_arg[i]).nonzero()
assert ranking.size(0) == 1
ranking = 1 + ranking.item()
type_relation = types_relations[
sample[:, 1][i].item()
]
# ranking + 1 is the true ranking used in evaluation metrics
metrics[mode][type_relation]['MRR'].update(1.0/ranking)
metrics[mode][type_relation]['MR'].update(ranking)
metrics[mode][type_relation]['HITS@1'].update(
1.0 if ranking <= 1 else 0.0)
metrics[mode][type_relation]['HITS@3'].update(
1.0 if ranking <= 3 else 0.0)
metrics[mode][type_relation]['HITS@10'].update(
1.0 if ranking <= 10 else 0.0)
if training:
model = model.train()
return metrics
def types_relations(self, model, dataset, threshold=1.5):
"""
Divide input dataset relations into different categories (i.e. ONE-TO-ONE, ONE-TO-MANY,
MANY-TO-ONE and MANY-TO-MANY) according to the mapping properties of relationships.
"""
stat_df = pd.DataFrame(self.true_triples)
stat_df.columns = ['head', 'relation', 'tail']
mean_head = stat_df[['head', 'relation', 'tail']].groupby(
['tail', 'relation']).count().groupby('relation').mean()
mean_tail = stat_df[['head', 'relation', 'tail']].groupby(
['head', 'relation']).count().groupby('relation').mean()
mean_relations = pd.concat(
[mean_head, mean_tail], axis='columns').reset_index()
mean_relations['head'] = mean_relations['head'].apply(
lambda x: '1' if x <= threshold else 'M')
mean_relations['tail'] = mean_relations['tail'].apply(
lambda x: '1' if x <= threshold else 'M')
mean_relations['type'] = mean_relations['head'] + \
'_' + mean_relations['tail']
mapping_type_relations = mean_relations.to_dict()['type']
relations_id = {value: key for key, value in self.relations.items()}
return {
relations_id[key]: value for key, value in mapping_type_relations.items()
}
def detail_eval(self, model, dataset, threshold=1.5):
"""
Divide input dataset relations into different categories (i.e. ONE-TO-ONE, ONE-TO-MANY,
MANY-TO-ONE and MANY-TO-MANY) according to the mapping properties of relationships.
Reference:
1. [Bordes, Antoine, et al. "Translating embeddings for modeling multi-relational data." Advances in neural information processing systems. 2013.](http://papers.nips.cc/paper/5071-translating-embeddings-for-modeling-multi-relational-data.pdf)
"""
mapping_type_relations = self.types_relations(
model=model,
dataset=dataset,
threshold=threshold
)
mapping_type_relations = {
self.relations[key]: value for key, value in mapping_type_relations.items()
}
types_relations = ['1_1', '1_M', 'M_1', 'M_M']
metrics = collections.OrderedDict({
'head-batch': collections.OrderedDict({}),
'tail-batch': collections.OrderedDict({})
})
for mode in ['head-batch', 'tail-batch']:
for type_relation in types_relations:
metrics[mode][type_relation] = collections.OrderedDict({
f'{metric}': stats.Mean()
for metric in ['MRR', 'MR', 'HITS@1', 'HITS@3', 'HITS@10']
})
with torch.no_grad():
for test_set in self.get_entity_stream(dataset):
metrics = self.compute_detailled_score(
model=model,
test_set=test_set,
metrics=metrics,
types_relations=mapping_type_relations,
device=self.device
)
for mode in ['head-batch', 'tail-batch']:
for type_relation in types_relations:
for metric in ['MRR', 'MR', 'HITS@1', 'HITS@3', 'HITS@10']:
metrics[mode][type_relation][metric] = round(
metrics[mode][type_relation][metric].get(), 4)
results = pd.DataFrame(metrics)
head = pd.DataFrame(results['head-batch'].values.tolist())
tail = pd.DataFrame(results['tail-batch'].values.tolist())
head.columns = pd.MultiIndex.from_product([["head"], head.columns])
tail.columns = pd.MultiIndex.from_product([["tail"], tail.columns])
results = pd.concat([head, tail], axis='columns')
results = results.set_index(pd.Series(['1_1', '1_M', 'M_1', 'M_M']))
results.index.name = 'relation'
# Add frequency of each type of relation:
frequency = collections.OrderedDict()
for type_relation in types_relations:
frequency[type_relation] = 0
for _, type_relation in mapping_type_relations.items():
frequency[type_relation] += 1
for type_relation in types_relations:
frequency[type_relation] /= len(mapping_type_relations)
frequency = pd.DataFrame.from_dict(
frequency,
orient='index',
columns=['frequency']
)
frequency.columns = pd.MultiIndex.from_product(
[["metadata"], frequency.columns]
)
results = pd.concat([results, frequency], axis='columns')
return results
| [
"raphael.sourty@gmail.com"
] | raphael.sourty@gmail.com |
7312d5425046c75f8b0c000b41197500ca3427cf | cbfaff10128807a84a91dfb52b720781bc07c29b | /airflow/dags/polygonetl_airflow/build_parse_dag.py | 4eb599cffa79ff6e4a4d8d3715a5496109d1fc11 | [
"MIT"
] | permissive | hansolavhilsen/polygon-etl | 7128516029bac7e112fe79c54e0508c086fa0f47 | ba45f5996a336f39864f6d29fbb8d7d7b40e76bb | refs/heads/main | 2023-07-03T11:25:39.039512 | 2021-08-09T11:21:47 | 2021-08-09T11:21:47 | 393,390,827 | 0 | 0 | MIT | 2021-08-06T13:48:54 | 2021-08-06T13:48:53 | null | UTF-8 | Python | false | false | 8,860 | py | from __future__ import print_function
import collections
import logging
import os
from datetime import datetime, timedelta
from glob import glob
from airflow import models
from airflow.operators.bash_operator import BashOperator
from airflow.operators.email_operator import EmailOperator
from airflow.operators.python_operator import PythonOperator
from airflow.operators.sensors import ExternalTaskSensor
from google.cloud import bigquery
from polygonetl_airflow.bigquery_utils import create_view
from polygonetl_airflow.common import read_json_file, read_file
from polygonetl_airflow.parse.parse_logic import ref_regex, parse, create_dataset
logging.basicConfig()
logging.getLogger().setLevel(logging.DEBUG)
dags_folder = os.environ.get('DAGS_FOLDER', '/home/airflow/gcs/dags')
def build_parse_dag(
dag_id,
dataset_folder,
parse_destination_dataset_project_id,
notification_emails=None,
parse_start_date=datetime(2020, 5, 30),
schedule_interval='0 0 * * *',
parse_all_partitions=None,
send_success_email=False
):
logging.info('parse_all_partitions is {}'.format(parse_all_partitions))
if parse_all_partitions:
dag_id = dag_id + '_FULL'
SOURCE_PROJECT_ID = 'public-data-finance'
SOURCE_DATASET_NAME = 'crypto_polygon'
PARTITION_DAG_ID = 'polygon_partition_dag'
default_dag_args = {
'depends_on_past': False,
'start_date': parse_start_date,
'email_on_failure': True,
'email_on_retry': False,
'retries': 5,
'retry_delay': timedelta(minutes=5)
}
if notification_emails and len(notification_emails) > 0:
default_dag_args['email'] = [email.strip() for email in notification_emails.split(',')]
dag = models.DAG(
dag_id,
catchup=False,
schedule_interval=schedule_interval,
default_args=default_dag_args)
validation_error = None
try:
validate_definition_files(dataset_folder)
except ValueError as e:
validation_error = e
# This prevents failing all dags as they are constructed in a loop in ethereum_parse_dag.py
if validation_error is not None:
def raise_validation_error(ds, **kwargs):
raise validation_error
validation_error_operator = PythonOperator(
task_id='validation_error',
python_callable=raise_validation_error,
provide_context=True,
execution_timeout=timedelta(minutes=10),
dag=dag
)
return dag
def create_parse_task(table_definition):
def parse_task(ds, **kwargs):
client = bigquery.Client()
parse(
bigquery_client=client,
table_definition=table_definition,
ds=ds,
source_project_id=SOURCE_PROJECT_ID,
source_dataset_name=SOURCE_DATASET_NAME,
destination_project_id=parse_destination_dataset_project_id,
sqls_folder=os.path.join(dags_folder, 'resources/stages/parse/sqls'),
parse_all_partitions=parse_all_partitions
)
table_name = table_definition['table']['table_name']
parsing_operator = PythonOperator(
task_id=table_name,
python_callable=parse_task,
provide_context=True,
execution_timeout=timedelta(minutes=60),
dag=dag
)
contract_address = table_definition['parser']['contract_address']
if contract_address is not None:
ref_dependencies = ref_regex.findall(table_definition['parser']['contract_address'])
else:
ref_dependencies = []
return parsing_operator, ref_dependencies
def create_add_view_task(dataset_name, view_name, sql):
def create_view_task(ds, **kwargs):
client = bigquery.Client()
dest_table_name = view_name
dest_table_ref = create_dataset(client, dataset_name, parse_destination_dataset_project_id).table(dest_table_name)
print('View sql: \n' + sql)
create_view(client, sql, dest_table_ref)
create_view_operator = PythonOperator(
task_id=f'create_view_{view_name}',
python_callable=create_view_task,
provide_context=True,
execution_timeout=timedelta(minutes=10),
dag=dag
)
return create_view_operator
wait_for_ethereum_load_dag_task = ExternalTaskSensor(
task_id='wait_for_polygon_partition_dag',
external_dag_id=PARTITION_DAG_ID,
external_task_id='done',
execution_delta=timedelta(minutes=30),
priority_weight=0,
mode='reschedule',
poke_interval=5 * 60,
timeout=60 * 60 * 12,
dag=dag)
json_files = get_list_of_files(dataset_folder, '*.json')
logging.info(json_files)
all_parse_tasks = {}
task_dependencies = {}
for json_file in json_files:
table_definition = read_json_file(json_file)
task, dependencies = create_parse_task(table_definition)
wait_for_ethereum_load_dag_task >> task
all_parse_tasks[task.task_id] = task
task_dependencies[task.task_id] = dependencies
checkpoint_task = BashOperator(
task_id='parse_all_checkpoint',
bash_command='echo parse_all_checkpoint',
priority_weight=1000,
dag=dag
)
for task, dependencies in task_dependencies.items():
for dependency in dependencies:
if dependency not in all_parse_tasks:
raise ValueError(
'Table {} is not found in the the dataset. Check your ref() in contract_address field.'.format(
dependency))
all_parse_tasks[dependency] >> all_parse_tasks[task]
all_parse_tasks[task] >> checkpoint_task
final_tasks = [checkpoint_task]
sql_files = get_list_of_files(dataset_folder, '*.sql')
logging.info(sql_files)
# TODO: Use folder name as dataset name and remove dataset_name in JSON definitions.
dataset_name = os.path.basename(dataset_folder)
full_dataset_name = 'polygon_' + dataset_name
for sql_file in sql_files:
sql = read_file(sql_file)
base_name = os.path.basename(sql_file)
view_name = os.path.splitext(base_name)[0]
create_view_task = create_add_view_task(full_dataset_name, view_name, sql)
checkpoint_task >> create_view_task
final_tasks.append(create_view_task)
if notification_emails and len(notification_emails) > 0 and send_success_email:
send_email_task = EmailOperator(
task_id='send_email',
to=[email.strip() for email in notification_emails.split(',')],
subject='Polygon ETL Airflow Parse DAG Succeeded',
html_content='Polygon ETL Airflow Parse DAG Succeeded for {}'.format(dag_id),
dag=dag
)
for final_task in final_tasks:
final_task >> send_email_task
return dag
def get_list_of_files(dataset_folder, filter='*.json'):
logging.info('get_list_of_files')
logging.info(dataset_folder)
logging.info(os.path.join(dataset_folder, filter))
return [f for f in glob(os.path.join(dataset_folder, filter))]
def validate_definition_files(dataset_folder):
json_files = get_list_of_files(dataset_folder, '*.json')
dataset_folder_name = dataset_folder.split('/')[-1]
all_lowercase_table_names = []
for json_file in json_files:
file_name = json_file.split('/')[-1].replace('.json', '')
table_definition = read_json_file(json_file)
table = table_definition.get('table')
if not table:
raise ValueError(f'table is empty in file {json_file}')
dataset_name = table.get('dataset_name')
if not dataset_name:
raise ValueError(f'dataset_name is empty in file {json_file}')
if dataset_folder_name != dataset_name:
raise ValueError(f'dataset_name {dataset_name} is not equal to dataset_folder_name {dataset_folder_name}')
table_name = table.get('table_name')
if not table_name:
raise ValueError(f'table_name is empty in file {json_file}')
if file_name != table_name:
raise ValueError(f'file_name {file_name} doest match the table_name {table_name}')
all_lowercase_table_names.append(table_name.lower())
table_name_counts = collections.defaultdict(lambda: 0)
for table_name in all_lowercase_table_names:
table_name_counts[table_name] += 1
non_unique_table_names = [name for name, count in table_name_counts.items() if count > 1]
if len(non_unique_table_names) > 0:
raise ValueError(f'The following table names are not unique {",".join(non_unique_table_names)}') | [
"araa@connect.ust.hk"
] | araa@connect.ust.hk |
f3d46fb64540cba1c12c46e08db3cba2eb1aa2f2 | f60f3ca2ea77cfaf2d12b949241252c42a0950c8 | /src/simulator/workflows/container.py | bceec745f29237b65de38323661bcadd15b7e136 | [] | no_license | belyakov-am/cloud-simulator | 39b1d84f4f0edba1098bd8d85f35d08106d4db04 | d2516adab27888a21e60651d536821c0a61092d8 | refs/heads/master | 2023-04-28T10:49:49.048826 | 2021-05-23T14:49:15 | 2021-05-23T14:49:15 | 344,893,767 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 736 | py | class Container:
"""Representation of (Docker) container with required libraries for
workflow execution.
"""
def __init__(self, workflow_uuid: str, provision_time: int) -> None:
self.workflow_uuid = workflow_uuid
self.provision_time = provision_time
def __repr__(self) -> str:
return (f"Container("
f"workflow_uuid = {self.workflow_uuid}, "
f"provision_time = {self.provision_time})")
def __eq__(self, other: "Container") -> bool:
return (self.workflow_uuid == other.workflow_uuid
and self.provision_time == other.provision_time)
def __hash__(self) -> int:
return hash(self.workflow_uuid) ^ hash(self.provision_time)
| [
"pochtaforfkn@gmail.com"
] | pochtaforfkn@gmail.com |
ab2312766b10746a33edee87aae7a0185bc0508e | 70ce903a7b835e4e960abe405158513790d37426 | /django-bloggy/bloggy_project/blog/models.py | 6e45b65afb50888a69149b4da6bd875560586d7b | [] | no_license | lpatmo/book2-exercises | 29af718d74732a5bbe287ab60a67b0d84d4e0abd | 9524bc58997ff4eda10177abf70805f3691e247c | refs/heads/master | 2020-12-25T22:29:09.391501 | 2014-10-26T03:15:35 | 2014-10-26T03:15:35 | 25,755,186 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 667 | py | from django.db import models
from uuslug import uuslug
class Post(models.Model):
created_at = models.DateTimeField(auto_now_add=True)
title = models.CharField(max_length=100)
content = models.TextField()
tag = models.CharField(max_length=20, blank=True, null=True)
image = models.ImageField(upload_to="images", blank=True, null=True)
views = models.IntegerField(default=0)
slug = models.CharField(max_length=100, unique=True)
def __unicode__(self):
return self.title
def save(self, *args, **kwargs):
self.slug = uuslug(self.title, instance=self, max_length=100)
super(Post, self).save(*args, **kwargs)
| [
"hermanmu@gmail.com"
] | hermanmu@gmail.com |
0d838535c8dd3ab286128f9df2c55d083a319f7d | cec8af4a2e2459d92db3b217ef729dcf0779c2dd | /website/urls.py | 0f382aa400a3d7654b0ac42e77a7c55ce7b38087 | [] | no_license | aashishnepal/pi-project | 1ee5db0345b42718ea172f2e6bd26f102b951c81 | dc76136d2bec2468537b75409fed121e3e10905a | refs/heads/master | 2020-11-25T18:32:36.785369 | 2019-12-27T08:49:52 | 2019-12-27T08:49:52 | 228,794,508 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 257 | py | from django.urls import path
from . import views
urlpatterns = [
path('', views.LoginPageView.as_view(), name='login'),
path('home/', views.HomePageView.as_view(), name='home'),
path('about/', views.AboutPageView.as_view(), name='about'),
]
| [
"aashishnepal008@gmail.com"
] | aashishnepal008@gmail.com |
bc9f514336676ddc97d1fc533448cc2a989f0a2d | 7dcbd6f01d33429925918afd3a7daff04e5b8346 | /rust_webrcon/oxide_commands.py | 6cb48c0a80a748d41c029cc7204d92a9602bad64 | [
"MIT"
] | permissive | thegreatstorm/rust_webrcon | 25514e92499672f462840c0711f11e7962bad1b0 | 2d0b670c57b30cceb5386334ed31e97e1a1f234e | refs/heads/master | 2022-11-30T01:42:26.948539 | 2020-08-13T23:05:14 | 2020-08-13T23:05:14 | 286,509,793 | 4 | 1 | null | 2020-08-10T23:30:22 | 2020-08-10T15:19:09 | Python | UTF-8 | Python | false | false | 2,951 | py | from rust_webrcon.utils.connection import connect_rust_rcon
def oxide_version(server_info):
command = 'oxide.version'
response = connect_rust_rcon(server_info, command)
return response
def oxide_plugin_list(server_info):
command = 'oxide.version'
response = connect_rust_rcon(server_info, command)
return response
def oxide_get_user_info(server_info, user):
command = 'oxide.show user {}'.format(user)
response = connect_rust_rcon(server_info, command)
return response
def oxide_get_group_info(server_info, group):
command = 'oxide.show group {}'.format(group)
response = connect_rust_rcon(server_info, command)
return response
def oxide_create_group(server_info, group):
command = 'oxide.group add {}'.format(group)
response = connect_rust_rcon(server_info, command)
return response
def oxide_remove_group(server_info, group):
command = 'oxide.group remove {}'.format(group)
response = connect_rust_rcon(server_info, command)
return response
def oxide_add_user_group(server_info, steam_id, group):
command = 'oxide.usergroup add {} {}'.format(steam_id,group)
response = connect_rust_rcon(server_info, command)
return response
def oxide_remove_user_group(server_info, steam_id, group):
command = 'oxide.usergroup remove {} {}'.format(steam_id,group)
response = connect_rust_rcon(server_info, command)
return response
def oxide_load_plugin(server_info, plugin_name):
command = 'oxide.load {}'.format(plugin_name)
response = connect_rust_rcon(server_info, command)
return response
def oxide_unload_plugin(server_info, plugin_name):
command = 'oxide.unload {}'.format(plugin_name)
response = connect_rust_rcon(server_info, command)
return response
def oxide_reload_plugin(server_info, plugin_name):
command = 'oxide.reload {}'.format(plugin_name)
response = connect_rust_rcon(server_info, command)
return response
def oxide_reload_all_plugins(server_info):
command = 'oxide.reload *'
response = connect_rust_rcon(server_info, command)
return response
def oxide_grant_user_perm(server_info, steam_id, permission):
command = 'oxide.grant user {} {}'.format(steam_id, permission)
response = connect_rust_rcon(server_info, command)
return response
def oxide_revoke_user_perm(server_info, steam_id, permission):
command = 'oxide.revoke user {} {}'.format(steam_id, permission)
response = connect_rust_rcon(server_info, command)
return response
def oxide_grant_group_perm(server_info, steam_id, permission):
command = 'oxide.grant group {} {}'.format(steam_id, permission)
response = connect_rust_rcon(server_info, command)
return response
def oxide_revoke_group_perm(server_info, steam_id, permission):
command = 'oxide.revoke group {} {}'.format(steam_id, permission)
response = connect_rust_rcon(server_info, command)
return response | [
"djasso@splunk.com"
] | djasso@splunk.com |
a5abed5427045f5e23f936aa700050d060c48ab1 | f5b340223a3dbde19889aaef01b9b18b6aab429a | /tasks_request.py | 3a2620d4c2131971893b8621a2e9e5b38cbf1eb0 | [] | no_license | jtpolo14/T0002 | a4b0399ce418d1607d2e4efa6e27f90eb12444bf | f099b62d92835bba600ce12d4a2a7d32a01156f6 | refs/heads/master | 2021-01-10T09:28:54.746517 | 2016-12-17T23:54:40 | 2016-12-17T23:54:40 | 53,239,581 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,286 | py | from celery import Celery
import time
from urllib.request import urlopen
import uuid
app = Celery('tasks', broker='redis://35.165.238.44', backend='redis://35.165.238.44')
@app.task
def check_input_file(file_path):
return file_path
@app.task
def move_input_file(file_path_start, file_path_end):
return (file_path_start, file_path_end)
@app.task
def word_count_python(infile='readme.txt'):
start = time.time()
with open(infile) as reader:
word_count = 0
for line in reader:
word_count += len(line.split())
time.sleep(5)
return (word_count, time.time() - start)
@app.task
def a001_get_file(file_url):
start = time.time()
ret = get_from_url(file_url)
if not ret['status'] == 0:
return ('a001 - error downloading url', time.time() - start)
else:
return (ret['file'], time.time() - start)
def get_from_url(url, prefix=None):
file_name = get_unique_file_name(prefix)
response = urlopen(url)
CHUNK = 16 * 1024
with open(file_name, 'wb') as f:
while True:
chunk = response.read(CHUNK)
if not chunk:
break
f.write(chunk)
return {'file':file_name, 'status':0}
def get_unique_file_name(prefix=None):
if prefix and type(prefix) == str:
return prefix + str(uuid.uuid4())
else:
return str(uuid.uuid4())
| [
"noreply@github.com"
] | noreply@github.com |
f0ccc3a5a140f298542c4fcf90576bdb112694b0 | e1fb0c63140a4cfebbe1c72fc2e5e76202e1b0f0 | /niftidataset/transforms.py | 32d4ac7829bdd7510a3daec502d659cded1be58a | [
"Apache-2.0"
] | permissive | s-mostafa-a/niftidataset | d11c1036d64fc5d25952fda5bfbd7a78d1507b0f | fa0175cf08753e6294d04e259b9b447a6dd3fa88 | refs/heads/master | 2022-11-15T18:36:10.118969 | 2020-07-11T09:53:50 | 2020-07-11T09:53:50 | 277,276,675 | 0 | 0 | NOASSERTION | 2020-07-05T10:16:05 | 2020-07-05T10:16:04 | null | UTF-8 | Python | false | false | 27,230 | py | #!/usr/bin/env python
# -*- coding: utf-8 -*-
"""
niftidataset.transforms
transformations to apply to images in dataset
Author: Jacob Reinhold (jacob.reinhold@jhu.edu)
Created on: Oct 24, 2018
"""
__all__ = ['RandomCrop2D',
'RandomCrop3D',
'RandomCrop',
'RandomSlice',
'ToTensor',
'ToPILImage',
'AddChannel',
'FixIntensityRange',
'Normalize',
'Digitize',
'MedianFilter',
'RandomAffine',
'RandomBlock',
'RandomFlip',
'RandomGamma',
'RandomNoise',
'TrimIntensity',
'get_transforms']
import random
from typing import Optional, Tuple, Union
import numpy as np
from PIL import Image
import torch
import torchvision as tv
import torchvision.transforms.functional as TF
from .errors import NiftiDatasetError
PILImage = type(Image)
class BaseTransform:
def __repr__(self): return f'{self.__class__.__name__}'
class CropBase(BaseTransform):
""" base class for crop transform """
def __init__(self, out_dim:int, output_size:Union[tuple,int,list], threshold:Optional[float]=None,
pct:Tuple[float,float]=(0.,1.), axis=0):
""" provide the common functionality for RandomCrop2D and RandomCrop3D """
assert isinstance(output_size, (int, tuple, list))
if isinstance(output_size, int):
self.output_size = (output_size,)
for _ in range(out_dim - 1):
self.output_size += (output_size,)
else:
assert len(output_size) == out_dim
self.output_size = output_size
self.out_dim = out_dim
self.thresh = threshold
self.pct = pct
self.axis = axis
def _get_sample_idxs(self, img: np.ndarray) -> Tuple[int, int, int]:
""" get the set of indices from which to sample (foreground) """
mask = np.where(img >= (img.mean() if self.thresh is None else self.thresh)) # returns a tuple of length 3
c = np.random.randint(0, len(mask[0])) # choose the set of idxs to use
h, w, d = [m[c] for m in mask] # pull out the chosen idxs
return h, w, d
def _offset_by_pct(self, h, w, d):
s = (h, w, d)
hml = wml = dml = 0
hmh = wmh = dmh = 0
i0, i1 = int(s[self.axis] * self.pct[0]), int(s[self.axis] * (1. - self.pct[1]))
if self.axis == 0:
hml += i0
hmh += i1
elif self.axis == 1:
wml += i0
wmh += i1
else:
dml += i0
dmh += i1
return (hml, wml, dml), (hmh, wmh, dmh)
def __repr__(self):
s = '{name}(output_size={output_size}, threshold={thresh})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomCrop2D(CropBase):
"""
Randomly crop a 2d slice/patch from a 3d image
Args:
output_size (tuple or int): Desired output size.
If int, cube crop is made.
axis (int or None): along which axis should the patch/slice be extracted
provide None for random axis
include_neighbors (bool): extract 3 neighboring slices instead of just 1
"""
def __init__(self, output_size:Union[int,tuple,list], axis:Optional[int]=0,
include_neighbors:bool=False, threshold:Optional[float]=None) -> None:
if axis is not None:
assert 0 <= axis <= 2
super().__init__(2, output_size, threshold)
self.axis = axis
self.include_neighbors = include_neighbors
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
axis = self.axis if self.axis is not None else np.random.randint(0, 3)
src, tgt = sample
*cs, h, w, d = src.shape
*ct, _, _, _ = src.shape
new_h, new_w = self.output_size
max_idxs = (np.inf, w - new_h//2, d - new_w//2) if axis == 0 else \
(h - new_h//2, np.inf, d - new_w//2) if axis == 1 else \
(h - new_h//2, w - new_w//2, np.inf)
min_idxs = (-np.inf, new_h//2, new_w//2) if axis == 0 else \
(new_h//2, -np.inf, new_w//2) if axis == 1 else \
(new_h//2, new_w//2, -np.inf)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
s_idxs = super()._get_sample_idxs(s)
idxs = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
s = self._get_slice(src, idxs, axis).squeeze()
t = self._get_slice(tgt, idxs, axis).squeeze()
if len(cs) == 0 or s.ndim == 2: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0 or t.ndim == 2: t = t[np.newaxis,...]
return s, t
def _get_slice(self, img:np.ndarray, idxs:Tuple[int,int,int], axis:int) -> np.ndarray:
h, w = self.output_size
n = 1 if self.include_neighbors else 0
oh = 0 if h % 2 == 0 else 1
ow = 0 if w % 2 == 0 else 1
i, j, k = idxs
s = img[..., i-n:i+1+n, j-h//2:j+h//2+oh, k-w//2:k+w//2+ow] if axis == 0 else \
img[..., i-h//2:i+h//2+oh, j-n:j+1+n, k-w//2:k+w//2+ow] if axis == 1 else \
img[..., i-h//2:i+h//2+oh, j-w//2:j+w//2+ow, k-n:k+1+n]
if self.include_neighbors:
s = np.transpose(s, (0,1,2)) if axis == 0 else \
np.transpose(s, (1,0,2)) if axis == 1 else \
np.transpose(s, (2,0,1))
return s
class RandomCrop3D(CropBase):
"""
Randomly crop a 3d patch from a (pair of) 3d image
Args:
output_size (tuple or int): Desired output size.
If int, cube crop is made.
"""
def __init__(self, output_size:Union[tuple,int,list], threshold:Optional[float]=None,
pct:Tuple[float,float]=(0.,1.), axis=0):
super().__init__(3, output_size, threshold, pct, axis)
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, h, w, d = src.shape
*ct, _, _, _ = tgt.shape
hh, ww, dd = self.output_size
(hml, wml, dml), (hmh, wmh, dmh) = self._offset_by_pct(h,w,d)
max_idxs = (h-hmh-hh//2, w-wmh-ww//2, d-dmh-dd//2)
min_idxs = (hml+hh//2, wml+ww//2, dml+dd//2)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
s_idxs = self._get_sample_idxs(s)
i, j, k = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
oh = 0 if hh % 2 == 0 else 1
ow = 0 if ww % 2 == 0 else 1
od = 0 if dd % 2 == 0 else 1
s = src[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow, k-dd//2:k+dd//2+od]
t = tgt[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow, k-dd//2:k+dd//2+od]
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
class RandomCrop:
"""
Randomly crop a 2d patch from a 2d image
Args:
output_size (tuple or int): Desired output size.
If int, square crop is made.
"""
def __init__(self, output_size:Union[tuple,int], threshold:Optional[float]=None):
self.output_size = (output_size, output_size) if isinstance(output_size, int) else output_size
self.thresh = threshold
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, h, w = src.shape
*ct, _, _ = tgt.shape
hh, ww = self.output_size
max_idxs = (h-hh//2, w-ww//2)
min_idxs = (hh//2, ww//2)
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
mask = np.where(s >= (s.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[0])) # choose the set of idxs to use
s_idxs = [m[c] for m in mask] # pull out the chosen idxs
i, j = [i if min_i <= i <= max_i else max_i if i > max_i else min_i
for max_i, min_i, i in zip(max_idxs, min_idxs, s_idxs)]
oh = 0 if hh % 2 == 0 else 1
ow = 0 if ww % 2 == 0 else 1
s = src[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow]
t = tgt[..., i-hh//2:i+hh//2+oh, j-ww//2:j+ww//2+ow]
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
def __repr__(self):
s = '{name}(output_size={output_size}, threshold={thresh})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomSlice(BaseTransform):
"""
take a random 2d slice from an image given a sample axis
Args:
axis (int): axis on which to take a slice
div (float): divide the mean by this value in the calculation of mask
the higher this value, the more background will be "valid"
"""
def __init__(self, axis:int=0, div:float=2):
assert 0 <= axis <= 2
self.axis = axis
self.div = div
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
src, tgt = sample
*cs, _, _, _ = src.shape
*ct, _, _, _ = tgt.shape
s = src[0] if len(cs) > 0 else src # use the first image to determine sampling if multimodal
idx = np.random.choice(self._valid_idxs(s)[self.axis])
s = self._get_slice(src, idx)
t = self._get_slice(tgt, idx)
if len(cs) == 0: s = s[np.newaxis,...] # add channel axis if empty
if len(ct) == 0: t = t[np.newaxis,...]
return s, t
def _get_slice(self, img:np.ndarray, idx:int):
s = img[...,idx,:,:] if self.axis == 0 else \
img[...,:,idx,:] if self.axis == 1 else \
img[...,:,:,idx]
return s
def _valid_idxs(self, img:np.ndarray) -> Tuple[np.ndarray,np.ndarray,np.ndarray]:
""" get the set of indices from which to sample (foreground) """
mask = np.where(img > img.mean() / self.div) # returns a tuple of length 3
h, w, d = [np.arange(np.min(m), np.max(m)+1) for m in mask] # pull out the valid idx ranges
return h, w, d
class ToTensor(BaseTransform):
""" Convert images in sample to Tensors """
def __init__(self, color=False):
self.color = color
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
if isinstance(src, np.ndarray) and isinstance(tgt, np.ndarray):
return torch.from_numpy(src), torch.from_numpy(tgt)
if isinstance(src, list): src = np.stack(src)
if isinstance(tgt, list): src = np.stack(tgt)
# handle PIL images
src, tgt = np.asarray(src), np.asarray(tgt)
if src.ndim == 3 and self.color: src = src.transpose((2,0,1)).astype(np.float32)
if tgt.ndim == 3 and self.color: tgt = tgt.transpose((2,0,1)).astype(np.float32)
if src.ndim == 2: src = src[None,...] # add channel dimension
if tgt.ndim == 2: tgt = tgt[None,...]
return torch.from_numpy(src), torch.from_numpy(tgt)
class ToPILImage(BaseTransform):
""" convert 2D image to PIL image """
def __init__(self, color=False):
self.color = color
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
src, tgt = np.squeeze(src), np.squeeze(tgt)
if src.ndim == 3 and self.color:
src = Image.fromarray(src.transpose((1,2,0)).astype(np.uint8))
elif src.ndim == 2:
src = Image.fromarray(src)
else:
src = [Image.fromarray(s) for s in src]
if tgt.ndim == 3 and self.color:
tgt = Image.fromarray(tgt.transpose((1,2,0)).astype(np.uint8))
elif tgt.ndim == 2:
tgt = Image.fromarray(tgt)
else:
tgt = [Image.fromarray(t) for t in tgt]
return src, tgt
class RandomAffine(tv.transforms.RandomAffine):
""" apply random affine transformations to a sample of images """
def __init__(self, p:float, degrees:float, translate:float=0, scale:float=0, resample:int=Image.BILINEAR,
segmentation=False):
self.p = p
self.degrees, self.translate, self.scale = (-degrees,degrees), (translate,translate), (1-scale,1+scale)
self.shear, self.fillcolor = None, 0
self.resample = resample
self.segmentation = segmentation
def affine(self, x, params, resample=Image.BILINEAR):
return TF.affine(x, *params, resample=resample, fillcolor=0)
def __call__(self, sample:Tuple[PILImage, PILImage]):
src, tgt = sample
ret = self.get_params(self.degrees, self.translate, self.scale, None, tgt.size)
if self.degrees[1] > 0 and random.random() < self.p:
if not isinstance(src, list):
src = self.affine(src, ret, self.resample)
else:
src = [self.affine(s, ret, self.resample) for s in src]
resample = Image.NEAREST if self.segmentation else self.resample
if not isinstance(tgt, list):
tgt = self.affine(tgt, ret, resample)
else:
tgt = [self.affine(t, ret, resample) for t in tgt]
return src, tgt
class RandomFlip:
def __init__(self, p:float, vflip:bool=False, hflip:bool=False):
self.p = p
self.vflip, self.hflip = vflip, hflip
def __call__(self, sample:Tuple[PILImage,PILImage]):
src, tgt = sample
if self.vflip and random.random() < self.p:
if not isinstance(src, list):
src = TF.vflip(src)
else:
src = [TF.vflip(s) for s in src]
if not isinstance(tgt, list):
tgt = TF.vflip(tgt)
else:
tgt = [TF.vflip(t) for t in tgt]
if self.hflip and random.random() < self.p:
if not isinstance(src, list):
src = TF.hflip(src)
else:
src = [TF.hflip(s) for s in src]
if not isinstance(tgt, list):
tgt = TF.hflip(tgt)
else:
tgt = [TF.hflip(t) for t in tgt]
return src, tgt
def __repr__(self):
s = '{name}(p={p}, vflip={vflip}, hflip={hflip})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomGamma:
""" apply random gamma transformations to a sample of images """
def __init__(self, p, tfm_y=False, gamma:float=0., gain:float=0.):
self.p, self.tfm_y = p, tfm_y
self.gamma, self.gain = (max(1-gamma,0),1+gamma), (max(1-gain,0),1+gain)
@staticmethod
def _make_pos(x): return x.min(), x - x.min()
def _gamma(self, x, gain, gamma):
is_pos = torch.all(x >= 0)
if not is_pos: m, x = self._make_pos(x)
x = gain * x ** gamma
if not is_pos: x = x + m
return x
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if random.random() < self.p:
gamma = random.uniform(self.gamma[0], self.gamma[1])
gain = random.uniform(self.gain[0], self.gain[1])
src = self._gamma(src, gain, gamma)
if self.tfm_y: tgt = self._gamma(tgt, gain, gamma)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, tfm_y={tfm_y}, gamma={gamma}, gain={gain})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomNoise:
""" add random gaussian noise to a sample of images """
def __init__(self, p, tfm_x=True, tfm_y=False, std:float=0):
self.p, self.tfm_x, self.tfm_y, self.std = p, tfm_x, tfm_y, std
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.std > 0 and random.random() < self.p:
if self.tfm_x: src = src + torch.randn_like(src).mul(self.std)
if self.tfm_y: tgt = tgt + torch.randn_like(tgt).mul(self.std)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, tfm_x={tfm_x}, tfm_y={tfm_y}, std={std})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class RandomBlock:
""" add random blocks of random intensity to a sample of images """
def __init__(self, p, sz_range, thresh=None, int_range=None, tfm_x=True, tfm_y=False, is_3d=False):
self.p, self.int, self.tfm_x, self.tfm_y = p, int_range, tfm_x, tfm_y
self.sz = sz_range if all([isinstance(szr, (tuple,list)) for szr in sz_range]) else \
(sz_range, sz_range, sz_range) if is_3d else (sz_range, sz_range)
self.thresh = thresh
self.is_3d = is_3d
def block2d(self, src, tgt):
_, hmax, wmax = src.shape
mask = np.where(src >= (src.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[1])) # choose the set of idxs to use
h, w = [m[c] for m in mask[1:]] # pull out the chosen idxs (2D)
sh, sw = random.randrange(*self.sz[0]), random.randrange(*self.sz[1])
oh = 0 if sh % 2 == 0 else 1
ow = 0 if sw % 2 == 0 else 1
if h+(sh//2)+oh >= hmax: h = hmax - (sh//2) - oh
if w+(sw//2)+ow >= wmax: w = wmax - (sw//2) - ow
if h-(sh//2) < 0: h = sh//2
if w-(sw//2) < 0: w = sw//2
int_range = self.int if self.int is not None else (src.min(), src.max()+1)
if random.random() < self.p:
if self.tfm_x: src[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow] = np.random.uniform(*int_range)
if self.tfm_y: tgt[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow] = np.random.uniform(*int_range)
return src, tgt
def block3d(self, src, tgt):
_, hmax, wmax, dmax = src.shape
mask = np.where(src >= (src.mean() if self.thresh is None else self.thresh))
c = np.random.randint(0, len(mask[1])) # choose the set of idxs to use
h, w, d = [m[c] for m in mask[1:]] # pull out the chosen idxs (2D)
sh, sw, sd = random.randrange(*self.sz[0]), random.randrange(*self.sz[1]), random.randrange(*self.sz[2])
oh = 0 if sh % 2 == 0 else 1
ow = 0 if sw % 2 == 0 else 1
od = 0 if sd % 2 == 0 else 1
if h+(sh//2)+oh >= hmax: h = hmax - (sh//2) - oh
if w+(sw//2)+ow >= wmax: w = wmax - (sw//2) - ow
if d+(sd//2)+od >= dmax: d = dmax - (sd//2) - od
if h-(sh//2) < 0: h = sh//2
if w-(sw//2) < 0: w = sw//2
if d-(sd//2) < 0: d = sd//2
int_range = self.int if self.int is not None else (src.min(), src.max()+1)
if isinstance(src, torch.Tensor): src, tgt = src.clone(), tgt.clone()
if random.random() < self.p:
if self.tfm_x: src[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow,d-sd//2:d+sd//2+od] = np.random.uniform(*int_range)
if self.tfm_y: tgt[:,h-sh//2:h+sh//2+oh,w-sw//2:w+sw//2+ow,d-sd//2:d+sd//2+od] = np.random.uniform(*int_range)
return src, tgt
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
src, tgt = self.block2d(src, tgt) if not self.is_3d else self.block3d(src, tgt)
return src, tgt
def __repr__(self):
s = '{name}(p={p}, sz={sz}, int_range={int}, thresh={thresh}, tfm_x={tfm_x}, tfm_y={tfm_y}, is_3d={is_3d})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class AddChannel:
""" Add empty first dimension to sample """
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
return (src.unsqueeze(0), tgt.unsqueeze(0))
class FixIntensityRange:
""" put data in range of 0 to 1 """
def __init__(self, scale:float=1):
self.scale = scale
def __call__(self, sample:Tuple[np.ndarray,np.ndarray]) -> Tuple[np.ndarray,np.ndarray]:
x, y = sample
x = self.scale * ((x - x.min()) / (x.max() - x.min()))
y = self.scale * ((y - y.min()) / (y.max() - y.min()))
return x, y
class Digitize:
""" digitize a sample of images """
def __init__(self, tfm_x=False, tfm_y=True, int_range=(1,100), step=1):
self.tfm_x, self.tfm_y, self.range, self.step = tfm_x, tfm_y, int_range, step
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = np.digitize(src, np.arange(self.range[0], self.range[1], self.step))
if self.tfm_y: tgt = np.digitize(tgt, np.arange(self.range[0], self.range[1], self.step))
return src, tgt
def normalize3d(tensor, mean, std, inplace=False):
"""
normalize a 3d tensor
Args:
tensor (Tensor): Tensor image of size (C, H, W, D) to be normalized.
mean (sequence): Sequence of means for each channel.
std (sequence): Sequence of standard deviations for each channel.
Returns:
Tensor: Normalized Tensor image.
"""
if not inplace:
tensor = tensor.clone()
mean = torch.as_tensor(mean, dtype=torch.float32, device=tensor.device)
std = torch.as_tensor(std, dtype=torch.float32, device=tensor.device)
tensor.sub_(mean[:, None, None, None]).div_(std[:, None, None, None])
return tensor
class Normalize:
"""
Implement a normalize function for input two images.
It computes std and mean for each input Tensor if mean and std equal to None,
then the function normalizes Tensor using the computed values.
Args:
mean: mean of input Tensor. if None passed, mean of each Tensor will be computed and normalization will be performed based on computed mean.
std: standard deviation of input Tensor. if None passed, std of each Tensor will be computed and normalization will be performed based on computed std.
tfm_x (bool): transform x or not
tfm_y (bool): transform y or not
is_3d (bool): is the Tensor 3d or not. this causes to normalize the Tensor on each channel.
"""
def __init__(self, mean=None, std=None, tfm_x:bool=True, tfm_y:bool=False,
is_3d:bool=False):
self.mean = mean
self.std = std
self.tfm_x = tfm_x
self.tfm_y = tfm_y
self.is_3d = is_3d
def _tfm(self, tensor:torch.Tensor):
if self.is_3d:
norm = normalize3d
mean = torch.as_tensor(self.mean, dtype=torch.float32, device=tensor.device) if not (
self.mean is None) else tensor.mean(dim=(1, 2, 3))
std = torch.as_tensor(self.std, dtype=torch.float32, device=tensor.device) if not (
self.std is None) else tensor.std(dim=(1, 2, 3))
# to prevent division by zero
std[std == 0.] = 1e-6
else:
norm = tv.transforms.functional.normalize
mean = self.mean if not (self.mean is None) else tensor.mean().item()
std = self.std if not (self.std is None) else tensor.std().item()
# to prevent division by zero
if std == 0.:
std = 1e-6
return norm(tensor, mean, std)
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = self._tfm(src)
if self.tfm_y: tgt = self._tfm(tgt)
return src, tgt
def __repr__(self):
s = '{name}(mean={mean}, std={std}, tfm_x={tfm_x}, tfm_y={tfm_y}, is_3d={is_3d})'
d = dict(self.__dict__)
return s.format(name=self.__class__.__name__, **d)
class MedianFilter:
""" median filter the sample """
def __init__(self, tfm_x=True, tfm_y=False):
try:
from scipy.ndimage.filters import median_filter
except (ModuleNotFoundError, ImportError):
raise NiftiDatasetError('scipy not installed, cannot use median filter')
self.filter = median_filter
self.tfm_x = tfm_x
self.tfm_y = tfm_y
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]):
src, tgt = sample
if self.tfm_x: src = self.filter(src, 3)
if self.tfm_y: tgt = self.filter(tgt, 3)
return src, tgt
class TrimIntensity:
"""
Trims intensity to given interval [new_min, new_max].
Trim intensities that are outside range [min_val, max_val], then scale to [new_min, new_max].
"""
def __init__(self, min_val:float, max_val:float,
new_min:float=-1.0, new_max:float=1.0, tfm_x:bool=True, tfm_y:bool=False):
if min_val >= max_val:
raise ValueError('min_val must be less than max_val')
if new_min >= new_max:
raise ValueError('new_min must be less than new_max')
self.min_val = min_val
self.max_val = max_val
self.new_min = new_min
self.new_max = new_max
self.tfm_x = tfm_x
self.tfm_y = tfm_y
def _tfm(self, x:torch.Tensor):
x = (x - self.min_val) / (self.max_val - self.min_val)
x[x > 1] = 1.
x[x < 0] = 0.
diff = self.new_max - self.new_min
x *= diff
x += self.new_min
return x
def __call__(self, sample:Tuple[torch.Tensor,torch.Tensor]) -> Tuple[torch.Tensor,torch.Tensor]:
src, tgt = sample
if self.tfm_x: src = self._tfm(src)
if self.tfm_y: tgt = self._tfm(tgt)
return src, tgt
def get_transforms(p:Union[list,float], tfm_x:bool=True, tfm_y:bool=False, degrees:float=0,
translate:float=None, scale:float=None, vflip:bool=False, hflip:bool=False,
gamma:float=0, gain:float=0, noise_pwr:float=0, block:Optional[Tuple[int,int]]=None,
thresh:Optional[float]=None, is_3d:bool=False,
mean:Optional[Tuple[float]]=None, std:Optional[Tuple[float]]=None,
color:bool=False, segmentation:bool=False):
""" get many desired transforms in a way s.t. can apply to nifti/tiffdatasets """
if isinstance(p, float): p = [p] * 5
tfms = []
do_affine = p[0] > 0 and (degrees > 0 or translate > 0 or scale > 0)
do_flip = p[1] > 0 and (vflip or hflip)
if do_affine or do_flip:
tfms.append(ToPILImage(color=color))
if do_affine:
tfms.append(RandomAffine(p[0], degrees, translate, scale, segmentation=segmentation))
if do_flip:
tfms.append(RandomFlip(p[1], vflip, hflip))
tfms.append(ToTensor(color))
if p[2] > 0 and (gamma is not None or gain is not None):
tfms.append(RandomGamma(p[2], tfm_y, gamma, gain))
if p[3] > 0 and (block is not None):
tfms.append(RandomBlock(p[3], block, thresh=thresh, tfm_x=tfm_x, tfm_y=tfm_y, is_3d=is_3d))
if p[4] > 0 and (noise_pwr > 0):
tfms.append(RandomNoise(p[4], tfm_x, tfm_y, noise_pwr))
if mean is not None and std is not None:
tfms.append(Normalize(mean, std, tfm_x, tfm_y, is_3d))
return tfms
| [
"jacob.reinhold@jhu.edu"
] | jacob.reinhold@jhu.edu |
1ab588834c32ac5271741da9e08ff13e07c2fdcf | 0c4b4c8c13f6ef1e26e20172e36cabd33851a17a | /radius_input_peri&cone.py | 5e428611e6005056ec0bc15b4ba985796f4c300e | [] | no_license | masankar666/Day1_26052021_pub | c1540c0660fd700daa190c8943b1e86af2de82ee | 312dc8206aefdb88172ff01cd860c766a6b1859f | refs/heads/main | 2023-06-04T10:32:43.362485 | 2021-06-26T14:28:23 | 2021-06-26T14:28:23 | 371,361,979 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 452 | py | '''
#Task 2
radius = 20
calculate perimeter of circle : 2 pi radius
'''
#dynamic radius input from user
r = float(input("Enter your radius as input:"))
h = float(input("Enter your height:"))
pi = 3.14
peri = 2 * pi * r
area = 1/3 * pi * (r ** 2) * h
print("Task 2 pgm 1 :")
print("radius = ",r)
print("height = ",h)
print ("calculate perimeter of circle = ",peri)
#area of cone
print ("area of cone = ",(int(area)))
| [
"noreply@github.com"
] | noreply@github.com |
1f9f53be7d85b393f7c0638c796d8ddc9f14b72f | 77090c3eaf15342505edc228ea19769ab219e0f7 | /CNVbenchmarkeR/output/manta3-datasetall/results17316/runWorkflow.py | 8ecfbc9ac2eb16a453983e3a063bca3a9ffd2a6b | [
"MIT"
] | permissive | robinwijngaard/TFM_code | 046c983a8eee7630de50753cff1b15ca3f7b1bd5 | d18b3e0b100cfb5bdd9c47c91b01718cc9e96232 | refs/heads/main | 2023-06-20T02:55:52.071899 | 2021-07-13T13:18:09 | 2021-07-13T13:18:09 | 345,280,544 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,090 | py | #!/usr/bin/env python2
# Workflow run script auto-generated by command: '/home/robin/Documents/Project/manta/Install/bin/configManta.py --bam=/home/robin/Documents/Project/Samples/bam/all/17316.bam --referenceFasta=/home/robin/Documents/Project/Samples/hg38/hg38.fa --config=/home/robin/Documents/Project/TFM_code/CNVbenchmarkeR/output/manta2-datasetall/configManta.py.ini --exome --runDir=/home/robin/Documents/Project/TFM_code/CNVbenchmarkeR/output/manta2-datasetall/results17316'
#
import os, sys
if sys.version_info >= (3,0):
import platform
raise Exception("Manta does not currently support python3 (version %s detected)" % (platform.python_version()))
if sys.version_info < (2,6):
import platform
raise Exception("Manta requires python2 version 2.6+ (version %s detected)" % (platform.python_version()))
scriptDir=os.path.abspath(os.path.dirname(__file__))
sys.path.append(r'/home/robin/Documents/Project/manta/Install/lib/python')
from mantaWorkflow import MantaWorkflow
def get_run_options(workflowClassName) :
from optparse import OptionGroup, SUPPRESS_HELP
from configBuildTimeInfo import workflowVersion
from configureUtil import EpilogOptionParser
from estimateHardware import EstException, getNodeHyperthreadCoreCount, getNodeMemMb
epilog="""Note this script can be re-run to continue the workflow run in case of interruption.
Also note that dryRun option has limited utility when task definition depends on upstream task
results -- in this case the dry run will not cover the full 'live' run task set."""
parser = EpilogOptionParser(description="Version: %s" % (workflowVersion), epilog=epilog, version=workflowVersion)
parser.add_option("-m", "--mode", type="string",dest="mode",
help=SUPPRESS_HELP)
parser.add_option("-j", "--jobs", type="string",dest="jobs",
help="number of jobs, must be an integer or 'unlimited' (default: Estimate total cores on this node)")
parser.add_option("-g","--memGb", type="string",dest="memGb",
help="gigabytes of memory available to run workflow, must be an integer (default: Estimate the total memory for this node)")
parser.add_option("-d","--dryRun", dest="isDryRun",action="store_true",default=False,
help="dryRun workflow code without actually running command-tasks")
parser.add_option("--quiet", dest="isQuiet",action="store_true",default=False,
help="Don't write any log output to stderr (but still write to workspace/pyflow.data/logs/pyflow_log.txt)")
def isLocalSmtp() :
import smtplib
try :
smtplib.SMTP('localhost')
except :
return False
return True
isEmail = isLocalSmtp()
emailHelp = SUPPRESS_HELP
if isEmail :
emailHelp="send email notification of job completion status to this address (may be provided multiple times for more than one email address)"
parser.add_option("-e","--mailTo", type="string",dest="mailTo",action="append",help=emailHelp)
debug_group = OptionGroup(parser,"development debug options")
debug_group.add_option("--rescore", dest="isRescore",action="store_true",default=False,
help="Reset task list to re-run hypothesis generation and scoring without resetting graph generation.")
parser.add_option_group(debug_group)
ext_group = OptionGroup(parser,"extended portability options (should not be needed by most users)")
ext_group.add_option("--maxTaskRuntime", type="string", metavar="hh:mm:ss",
help="Specify max runtime per task (no default)")
parser.add_option_group(ext_group)
(options,args) = parser.parse_args()
if not isEmail : options.mailTo = None
if len(args) :
parser.print_help()
sys.exit(2)
if options.mode is None :
options.mode = "local"
elif options.mode not in ["local"] :
parser.error("Invalid mode. Available modes are: local")
if options.jobs is None :
try :
options.jobs = getNodeHyperthreadCoreCount()
except EstException:
parser.error("Failed to estimate cores on this node. Please provide job count argument (-j).")
if options.jobs != "unlimited" :
options.jobs=int(options.jobs)
if options.jobs <= 0 :
parser.error("Jobs must be 'unlimited' or an integer greater than 1")
# note that the user sees gigs, but we set megs
if options.memGb is None :
try :
options.memMb = getNodeMemMb()
except EstException:
parser.error("Failed to estimate available memory on this node. Please provide available gigabyte argument (-g).")
elif options.memGb != "unlimited" :
options.memGb=int(options.memGb)
if options.memGb <= 0 :
parser.error("memGb must be 'unlimited' or an integer greater than 1")
options.memMb = 1024*options.memGb
else :
options.memMb = options.memGb
options.resetTasks=[]
if options.isRescore :
options.resetTasks.append("makeHyGenDir")
return options
def main(pickleConfigFile, primaryConfigSection, workflowClassName) :
from configureUtil import getConfigWithPrimaryOptions
runOptions=get_run_options(workflowClassName)
flowOptions,configSections=getConfigWithPrimaryOptions(pickleConfigFile,primaryConfigSection)
# new logs and marker files to assist automated workflow monitoring:
warningpath=os.path.join(flowOptions.runDir,"workflow.warning.log.txt")
errorpath=os.path.join(flowOptions.runDir,"workflow.error.log.txt")
exitpath=os.path.join(flowOptions.runDir,"workflow.exitcode.txt")
# the exit path should only exist once the workflow completes:
if os.path.exists(exitpath) :
if not os.path.isfile(exitpath) :
raise Exception("Unexpected filesystem item: '%s'" % (exitpath))
os.unlink(exitpath)
wflow = workflowClassName(flowOptions)
retval=1
try:
retval=wflow.run(mode=runOptions.mode,
nCores=runOptions.jobs,
memMb=runOptions.memMb,
dataDirRoot=flowOptions.workDir,
mailTo=runOptions.mailTo,
isContinue="Auto",
isForceContinue=True,
isDryRun=runOptions.isDryRun,
isQuiet=runOptions.isQuiet,
resetTasks=runOptions.resetTasks,
successMsg=wflow.getSuccessMessage(),
retryWindow=0,
retryMode='all',
warningLogFile=warningpath,
errorLogFile=errorpath)
finally:
exitfp=open(exitpath,"w")
exitfp.write("%i\n" % (retval))
exitfp.close()
sys.exit(retval)
main(r"/home/robin/Documents/Project/TFM_code/CNVbenchmarkeR/output/manta2-datasetall/results17316/runWorkflow.py.config.pickle","manta",MantaWorkflow)
| [
"robinwijngaard@gmail.com"
] | robinwijngaard@gmail.com |
95c6dc627b9d0477a4aace75cad812e43a612748 | abc6015c3c4bc7b15de81d8bdc920674bfd63cd6 | /TGNDA_fog.py | 160e9f1c131c685fd17c41a5c7fffd3987c47d48 | [] | no_license | tnda-har/tda4fog | 2ff7c33efd8d557ad6a095e3b6b52e28c815724c | 3849757da45308cbb63b9bd5e4fbd80ea2d17f5a | refs/heads/main | 2023-07-17T08:15:41.513970 | 2021-09-03T07:08:12 | 2021-09-03T07:08:12 | 368,104,512 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 8,906 | py | # %load test.py
import math
import os
import random
import time
from numpy import interp
import pandas as pd
import numpy as np
from gtda.diagrams import Scaler, BettiCurve, PersistenceLandscape, Silhouette
from gtda.homology import VietorisRipsPersistence
from gtda.pipeline import make_pipeline
from gtda.time_series import TakensEmbedding
from sklearn.ensemble import RandomForestClassifier
from sklearn.metrics import classification_report
from sklearn.model_selection import train_test_split
from sklearn import metrics
import matplotlib.pyplot as plt
from sklearn.model_selection import KFold
from sklearn.metrics import roc_curve, auc
import warnings
def get_random_list(start, stop, n):
arr = list(range(start, stop + 1))
shuffle_n(arr, n)
return arr[-n:]
def shuffle_n(arr, n):
random.seed(time.time())
for i in range(len(arr) - 1, len(arr) - n - 1, -1):
j = random.randint(0, i)
arr[i], arr[j] = arr[j], arr[i]
def create_window(act, window_length, dataframe):
indices = list(dataframe[dataframe.Action == act].index)
groups = []
temp = []
group_count = 0
for i in range(len(indices)):
if i == len(indices) - 1:
temp.append(indices[i])
groups.append(temp)
temp = []
break
temp.append(indices[i])
if indices[i] + 1 != indices[i + 1]:
group_count += 1
groups.append(temp)
temp = []
fs = 64
final_dataframe = pd.DataFrame()
for i in groups:
required = math.floor(len(i) / (window_length * fs))
req_index = i[0:(required * window_length * fs)]
final_dataframe = pd.concat([final_dataframe, dataframe.iloc[req_index, :]], axis=0)
return final_dataframe
def sbj_df(df, sbj='S01'):
DF0 = create_window(0, 2, df)
DF0 = DF0[DF0['name'] == sbj]
DF1 = create_window(1, 2, df)
DF1 = DF1[DF1['name'] == sbj]
DF2 = create_window(2, 2, df)
DF2 = DF2[DF2['name'] == sbj]
return DF0, DF1, DF2
def gtda(dataframe, w=128):
all_data = []
dataframe = dataframe.drop(columns=['time', 'Action', 'name'])
for i in range(0, len(dataframe), w):
data = dataframe.iloc[i:i + w]
data = data.to_numpy().transpose()
if data.shape[1] == w:
all_data.append(data)
all_data = np.array(all_data)
steps = [TakensEmbedding(time_delay=5, dimension=3),
VietorisRipsPersistence(),
Scaler()
]
tda_pipe = make_pipeline(*steps)
diagrams = tda_pipe.fit_transform(all_data)
BC = BettiCurve(n_bins=50).fit_transform(diagrams)
PL = PersistenceLandscape(n_bins=50).fit_transform(diagrams)
SL = Silhouette(n_bins=50).fit_transform(diagrams)
return np.mean(BC, axis=1), np.sum(PL, axis=1), np.mean(SL, axis=1)
def training(x, y, sbj_name):
y = np.array(y)
x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.3, shuffle=True)
model = RandomForestClassifier(random_state=1)
model.fit(x_train, y_train)
y_pre = model.predict(x_test)
print(classification_report(y_test, y_pre, digits=4))
pre_y = model.predict_proba(x_test)[:, 1]
fpr, tpr, threshold = metrics.roc_curve(y_test, pre_y)
roc_auc = metrics.auc(fpr, tpr)
print('AUC:', roc_auc)
print('\n')
cv = KFold(n_splits=5, shuffle=True, random_state=None)
tprs = []
aucs = []
mean_fpr = np.linspace(0, 1, 100)
i = 0
for train_index, test_index in cv.split(x):
X_train, X_test = x[train_index], x[test_index]
Y_train, Y_test = y[train_index], y[test_index]
build_model = model
build_model.fit(X_train, Y_train.astype("int"))
Y_pre = build_model.predict_proba(X_test)[:, 1]
fpr, tpr, thresholds = roc_curve(Y_test, Y_pre)
tprs.append(interp(mean_fpr, fpr, tpr))
tprs[-1][0] = 0.0
roc_auc = auc(fpr, tpr)
aucs.append(roc_auc)
plt.plot(fpr, tpr, lw=1, alpha=.6, label='ROC fold %d(AUC=%0.2f)' % (i, roc_auc))
i += 1
plt.plot([0, 1], [0, 1], linestyle='--', lw=2, color='r', label='Luck', alpha=.8)
mean_tpr = np.mean(tprs, axis=0)
mean_tpr[-1] = 1.0
mean_auc = auc(mean_fpr, mean_tpr) # 计算平均AUC值
plt.plot(mean_fpr, mean_tpr, color='b', label=r'Mean ROC (AUC=%0.2f)' % mean_auc, lw=2, alpha=.8)
std_tpr = np.std(tprs, axis=0)
tprs_upper = np.minimum(mean_tpr + std_tpr, 1)
tprs_lower = np.maximum(mean_tpr - std_tpr, 0)
print('mean auc', mean_auc)
plt.fill_between(mean_tpr, tprs_lower, tprs_upper, color='gray', alpha=.2)
plt.xlim([-0.05, 1.05])
plt.ylim([-0.05, 1.05])
plt.xlabel('False Positive Rate')
plt.ylabel('True Positive Rate')
plt.title(sbj_name + 'FOG_ROC')
plt.legend(loc='lower right')
plt.show()
return
def training_plot(x, y, m):
y = np.array(y)
x_train, x_test, y_train, y_test = train_test_split(x, y, test_size=0.3, shuffle=True)
model = RandomForestClassifier(random_state=1)
model.fit(x_train, y_train)
y_pre = model.predict(x_test)
print(classification_report(y_test, y_pre, digits=4))
pre_y = model.predict_proba(x_test)[:, 1]
fpr, tpr, threshold = metrics.roc_curve(y_test, pre_y)
roc_auc = metrics.auc(fpr, tpr)
print('AUC:', roc_auc)
print('\n')
cv = KFold(n_splits=5, shuffle=True, random_state=None)
tprs = []
aucs = []
mean_fpr = np.linspace(0, 1, 100)
i = 0
for train_index, test_index in cv.split(x):
X_train, X_test = x[train_index], x[test_index]
Y_train, Y_test = y[train_index], y[test_index]
build_model = model
build_model.fit(X_train, Y_train.astype("int"))
Y_pre = build_model.predict_proba(X_test)[:, 1]
fpr, tpr, thresholds = roc_curve(Y_test, Y_pre)
tprs.append(interp(mean_fpr, fpr, tpr))
tprs[-1][0] = 0.0
roc_auc = auc(fpr, tpr)
aucs.append(roc_auc)
i += 1
mean_tpr = np.mean(tprs, axis=0)
mean_tpr[-1] = 1.0
mean_auc = auc(mean_fpr, mean_tpr)
print('mean auc', mean_auc)
plt.plot(fpr, tpr, lw=1, alpha=.6, label='FOG AUC of %s = %0.3f' % (m, mean_auc))
def clean_nan(x):
x = np.nan_to_num(x, posinf=0, neginf=0)
return x
warnings.filterwarnings("ignore")
df = pd.read_csv('TGNDA_for_FOG2.csv')
subject_list = ['S01', 'S02', 'S03', 'S05', 'S06', 'S07', 'S08', 'S09']
whole_x = []
whole_y = []
whole_bc_x = []
whole_bc_y = []
whole_pl_x = []
whole_pl_y = []
whole_sl_x = []
whole_sl_y = []
for sbj in subject_list:
DF0, DF1, DF2 = sbj_df(df, sbj=sbj)
NOR_DF = pd.concat([DF0, DF2])
nor_bc, nor_pl, nor_sl = gtda(NOR_DF)
fog_bc, fog_pl, fog_sl = gtda(DF1)
print(len(fog_bc))
fog = [fog_bc, fog_pl, fog_sl]
normal = [nor_bc, nor_pl, nor_sl]
idx = get_random_list(0, len(fog_bc) - 1, int(len(fog_bc) * 1))
for length, nor_feature in enumerate(normal):
fog_feature = fog[length]
nor_feature = nor_feature[idx]
X = np.concatenate([nor_feature, fog_feature], axis=0)
Y = []
Y += len(nor_feature) * [0]
Y += len(fog_feature) * [1]
if length == 0:
whole_bc_x.append(X)
whole_bc_y += Y
elif length == 1:
whole_pl_x.append(X)
whole_pl_y += Y
elif length == 2:
whole_sl_x.append(X)
whole_sl_y += Y
print(sbj, '第{}个特征'.format(length + 1))
training(X, Y, sbj)
all_fog = np.concatenate(fog, axis=1)
all_nor = np.concatenate(normal, axis=1)
all_nor = all_nor[idx]
all_X = np.concatenate([all_nor, all_fog], axis=0)
all_Y = []
all_Y += len(all_nor) * [0]
all_Y += len(all_fog) * [1]
print(sbj, '特征融合')
training(all_X, all_Y, sbj)
whole_x.append(all_X)
whole_y += all_Y
whole_bc_x = np.concatenate(whole_bc_x, axis=0)
whole_pl_x = np.concatenate(whole_pl_x, axis=0)
whole_sl_x = np.concatenate(whole_sl_x, axis=0)
whole_x = np.concatenate(whole_x, axis=0)
whole = 'All Person'
plt.figure()
print('all person BC')
training_plot(whole_bc_x, whole_bc_y, 'BC')
print('all person PL')
training_plot(whole_pl_x, whole_pl_y, 'PL')
print('all person SL')
training_plot(whole_sl_x, whole_sl_y, 'SL')
print('all person whole')
training_plot(whole_x, whole_y, 'Fusion')
plt.title(whole + ' FOG ROC')
plt.legend(loc='lower right')
plt.plot([0, 1], [0, 1], linestyle='--', lw=2, color='r', label='Luck', alpha=.8)
plt.xlim([-0.05, 1.05])
plt.ylim([-0.05, 1.05])
plt.ylabel('True Positive Rate')
plt.xlabel('False Positive Rate')
plt.show()
| [
"noreply@github.com"
] | noreply@github.com |
42f319f1590604485a6f3d6007b2a3d6c014d924 | bfb5051ae8de82ec9ebfd329336e6aa6e7cea3d9 | /src/entities/difficulty.py | 4da58572326508ea2a53998c5c2206c083771a6d | [] | no_license | J-Uhero/ot-harjoitustyo | 400cadf132eaa78a5e106ab9ae409c804ef35745 | dc04dd58d910093ae4c9fb3ac581e3a99b72513b | refs/heads/master | 2023-04-29T07:55:47.081100 | 2021-05-19T23:39:33 | 2021-05-19T23:39:33 | 350,096,011 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,028 | py | class Difficulty:
"""Luokka, joka kuvaa pelin vaikeustasoa ja sisältää vaikeustasoon
vaikuttavia muuttujia ja niiden parametreja.
"""
def __init__(self):
"""Konstruktori, joka alustaa vaikeustasoa ilmentävät muuttujat
asettaa oletusarvoisesti vaikeustason helpoksi.
"""
self._height = None
self._width = None
self._mines = None
self._degree = None
self.easy()
def height(self):
return self._height
def width(self):
return self._width
def mines(self):
return self._mines
def degree(self):
return self._degree
def easy(self):
self._height = 9
self._width = 9
self._mines = 10
self._degree = "easy"
def medium(self):
self._height = 16
self._width = 16
self._mines = 40
self._degree = "medium"
def hard(self):
self._height = 16
self._width = 30
self._mines = 99
self._degree = "hard"
| [
"juho.herranen@gmail.com"
] | juho.herranen@gmail.com |
787ab026083a3a4bd73e4ea8f22df850b46410bd | b7e72138d624d21fbeb877d0dd5613389fb70503 | /rifaqui/core/migrations/0009_raffle_draw_date.py | c5592c5f1dc6a7f21ce2c8dda2c8a003361de9e7 | [] | no_license | ConTTudOweb/RifaquiProject | f4f2667e7185df517b8919013f505ccd68d8d901 | 301b4d881a1e820774f9a74359dcbd7eff3d7e45 | refs/heads/master | 2020-03-15T06:33:34.094190 | 2018-06-22T03:25:43 | 2018-06-22T03:25:43 | 132,010,143 | 0 | 2 | null | null | null | null | UTF-8 | Python | false | false | 424 | py | # Generated by Django 2.0.4 on 2018-05-02 21:32
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('core', '0008_auto_20180502_2122'),
]
operations = [
migrations.AddField(
model_name='raffle',
name='draw_date',
field=models.DateField(blank=True, null=True, verbose_name='data do sorteio'),
),
]
| [
"alessandro@kminformatica.com.br"
] | alessandro@kminformatica.com.br |
808afd2c166dd88286794b21c33a75891fcad75a | eb0bb5267035c0222da0c072c5dcd85b46099904 | /test/bug.986.t | 7d7e22538d6c69ad56a124722cd5c465bf5b6fda | [
"MIT"
] | permissive | bjornreppen/task | 6d96f578eec7b9cceeb4d728caeda87e7a446949 | a9eac8bb715ac8f51073c080ac439bf5c09493e8 | refs/heads/master | 2021-05-30T07:48:39.263967 | 2015-10-21T20:50:42 | 2015-10-21T20:50:42 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,329 | t | #!/usr/bin/env python2.7
# -*- coding: utf-8 -*-
###############################################################################
#
# Copyright 2006 - 2015, Paul Beckingham, Federico Hernandez.
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included
# in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.
#
# http://www.opensource.org/licenses/mit-license.php
#
###############################################################################
import sys
import os
import unittest
# Ensure python finds the local simpletap module
sys.path.append(os.path.dirname(os.path.abspath(__file__)))
from basetest import Task, TestCase
from basetest import Taskd, ServerTestCase
class TestBug986(TestCase):
def setUp(self):
"""Executed before each test in the class"""
self.t = Task()
def test_dateformat_precedence(self):
"""Verify rc.dateformat.info takes precedence over rc.dateformat"""
self.t('add test')
self.t('1 start')
code, out, err = self.t('1 info rc.dateformat:XX rc.dateformat.info:__')
self.assertIn('__', out)
self.assertNotIn('XX', out)
code, out, err = self.t('1 info rc.dateformat:__ rc.dateformat.info:')
self.assertIn('__', out)
if __name__ == "__main__":
from simpletap import TAPTestRunner
unittest.main(testRunner=TAPTestRunner())
# vim: ai sts=4 et sw=4 ft=python
| [
"paul@beckingham.net"
] | paul@beckingham.net |
c835bb059b3de560e44f8f5470abdf3e721dc9fe | 3d8c0fddfa020c3c29fc244bba5c89aa9e8d5f15 | /lab_5/naive_bayes_model.py | 51cab803438b30b81db702bd6c8c6e32c4aa88d1 | [] | no_license | Max-1892/machine_learning | 43c24ea03ccc26d67ef015447b7e5525401cc0bd | ec363911fc913e8271e629657ffb0a2377c9571f | refs/heads/master | 2020-01-27T10:04:41.455509 | 2019-07-07T18:49:13 | 2019-07-07T18:49:13 | 67,073,296 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,793 | py | import numpy as np
class NaiveBayes:
"""
This class encapsulates the naive Bayes algorithm.
Note: if a probability of 0 comes up during the prediction stage
a very small number is used for the calculation in it's place.
"""
def __init__(self):
self.class_priors = {0: 0, 1: 0}
# Represents table of conditional probabilities,
# first key is class value, second value is value of attributes in
# a particular row of the table
self.feature_given_class_prob = {0: {0: {}, 1: {}}, 1: {0: {}, 1: {}}}
def build_model(self, data_instances):
# Split on class
negative_instances = []
positive_instances = []
for instance in data_instances:
if instance[-1] == 0:
negative_instances.append(instance[:-1])
elif instance[-1] == 1:
positive_instances.append(instance[:-1])
# Calculate class priors
self.class_priors[0] = float(len(negative_instances)) / len(data_instances)
self.class_priors[1] = float(len(positive_instances)) / len(data_instances)
negative_instances = np.array(negative_instances)
positive_instances = np.array(positive_instances)
# Calculate feature given class
for class_value in range(2):
for feature_value in range(2):
for feature_idx in range(len(instance[:-1])):
if class_value == 0:
if len(negative_instances) > 0:
self.feature_given_class_prob[class_value][feature_value][feature_idx] = \
float((negative_instances[:,feature_idx] == feature_value).sum()) / len(negative_instances)
else:
self.feature_given_class_prob[class_value][feature_value][feature_idx] = 0.0
if class_value == 1:
if len(positive_instances) > 0:
self.feature_given_class_prob[class_value][feature_value][feature_idx] = \
float((positive_instances[:,feature_idx] == feature_value).sum()) / len(positive_instances)
else:
self.feature_given_class_prob[class_value][feature_value][feature_idx] = 0.0
def predict(self, data_instance):
argmax_class_probability = -1.0
argmax_class = -1
for class_value in range(2):
probability_of_belonging_to_class_value = self.class_priors[class_value]
for attr_idx, instance_attr in enumerate(data_instance):
if self.feature_given_class_prob[class_value][instance_attr][attr_idx] == 0:
probability_of_belonging_to_class_value *= 0.00000000000000000000001
else:
probability_of_belonging_to_class_value *= \
self.feature_given_class_prob[class_value][instance_attr][attr_idx]
if probability_of_belonging_to_class_value > argmax_class_probability:
argmax_class_probability = probability_of_belonging_to_class_value
argmax_class = class_value
return argmax_class
def print_model(self):
model = "P(class value = {}) = {}\n".format(0, self.class_priors[0])
model += "P(class value = {}) = {}\n".format(1, self.class_priors[1])
for class_key, inst_val_map in self.feature_given_class_prob.iteritems():
for instance_val_key, value in inst_val_map.iteritems():
for attr_key, probability in value.iteritems():
model += "P(f_{} = {} | Class value = {}) = {}\n".format( \
attr_key, instance_val_key, class_key, probability)
return model
| [
"somethingfunny14@gmail.com"
] | somethingfunny14@gmail.com |
0f56d4ad2504d3fc8c7bc698669c0f95e8b2b0c0 | ca7aa979e7059467e158830b76673f5b77a0f5a3 | /Python_codes/p02701/s368667639.py | 8111538afd640004a5f4d8e54edf3e76da2a6571 | [] | no_license | Aasthaengg/IBMdataset | 7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901 | f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8 | refs/heads/main | 2023-04-22T10:22:44.763102 | 2021-05-13T17:27:22 | 2021-05-13T17:27:22 | 367,112,348 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 43 | py | print(len(set(open(0).read().split())) - 1) | [
"66529651+Aastha2104@users.noreply.github.com"
] | 66529651+Aastha2104@users.noreply.github.com |
77bf35076a85c2afd353540ed90cac6cc4258fa7 | f1ab02aa2593867c695d290bd06153c14cf0f098 | /npf/core/xmin/util/admin.py | d37563c3504757c9badc8a1b05e42f3aa296c801 | [] | no_license | 1npf/EDISSON-FAZA-1 | 0d3fa9c4836ae0bd4b405de646f3220687506cc5 | 81d78a8a0df4ba88cb4a32c6d4a616ca9a9e6ae1 | refs/heads/master | 2016-09-01T16:23:59.131240 | 2015-12-24T12:16:55 | 2015-12-24T12:16:55 | 48,541,140 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,065 | py | from django.contrib import admin
from django.core.urlresolvers import reverse, NoReverseMatch
from django.db import models
from django.http import Http404
from django.utils.html import escape
from django.contrib.contenttypes.models import ContentType
from npf.core.printform.models import Template
try:
from django.utils.encoding import force_unicode
except ImportError:
from django.utils.encoding import force_text as force_unicode
def get_admin_for_model(model):
return get_model_and_admin(model._meta.app_label, model._meta.model_name)
def get_model_and_admin(app_label, model_name):
for (model, model_admin) in admin.site._registry.items():
if model._meta.app_label == app_label and model._meta.model_name == model_name:
return model, model_admin
return None
def get_model_and_admin_or_404(app_label, module_name):
model_and_admin = get_model_and_admin(app_label, module_name)
if model_and_admin is None:
raise Http404
return model_and_admin
def get_admin_urls_for_model(request, model):
admin_urls = {}
model_and_admin = get_admin_for_model(model)
if model_and_admin is None:
return admin_urls
model_admin = model_and_admin[1]
app_label = model._meta.app_label
has_module_perms = request.user.has_module_perms(app_label)
if has_module_perms:
perms = model_admin.get_model_perms(request)
info = (app_label, model._meta.model_name)
if perms.get('change', False):
try:
admin_urls['admin_url'] = reverse('admin:%s_%s_changelist' % info, current_app=admin.site.name)
except NoReverseMatch:
pass
if perms.get('add', False):
try:
admin_urls['add_url'] = reverse('admin:%s_%s_add' % info, current_app=admin.site.name)
except NoReverseMatch:
pass
return admin_urls
def flatten_choices(choices):
# Normalize to strings.
output = []
for option_value, option_label in choices:
if isinstance(option_label, (list, tuple)):
output.append({'id': '', 'type': 'header', 'description': escape(force_unicode(option_value))})
for option in option_label:
output.append({'id': option[0], 'type': 'item', 'description': escape(force_unicode(option[1]))})
else:
output.append({'id': option_value, 'type': 'header_item',
'description': escape(force_unicode(option_label))})
return output
def get_app_list(request):
def _actions_exists(model_admin, model):
if bool(model_admin.get_action_choices(request)):
return True
ct = ContentType.objects.get_for_model(model, not model._meta.proxy)
if Template.objects.filter(model=ct).exists():
return True
return False
app_dict = {}
user = request.user
for model, model_admin in admin.site._registry.items():
app_label = model._meta.app_label
app_name = model._meta.app_config.verbose_name
has_module_perms = user.has_module_perms(app_label)
if has_module_perms:
perms = model_admin.get_model_perms(request)
# Check whether user has any perm for this module.
# If so, add the module to the model_list.
if True in perms.values():
fields = []
for field in model._meta.fields:
field_config = {
'name': field.name,
'verbose_name': field.verbose_name,
'app': app_label,
'model': model._meta.model_name,
'field_class': "%s.%s" % (field.__class__.__module__, field.__class__.__name__),
'default': field.get_default(),
'editable': field.editable,
'allow_blank': field.blank,
'help_text': field.help_text,
'max_length': field.max_length,
'choices': flatten_choices(field.choices)
}
if isinstance(field, models.ForeignKey):
field_config['related'] = {'class': "%s.%s" % (field.rel.to.__module__, field.rel.to.__name__)}
field_config['related'].update(get_admin_urls_for_model(request, field.rel.to))
fields.append(field_config)
model_dict = {
'app': app_label,
'model': model._meta.model_name,
'model_name': model._meta.object_name,
'verbose_name': model._meta.verbose_name,
'verbose_name_plural': model._meta.verbose_name_plural,
'perms': perms,
'list_display': model_admin.get_list_display(request),
'list_editable': model_admin.list_editable,
'list_per_page': model_admin.list_per_page,
'search_fields': model_admin.get_search_fields(request),
'fields': fields,
'actions': _actions_exists(model_admin, model)
}
model_dict.update(get_admin_urls_for_model(request, model))
if hasattr(model_admin, 'columns'):
model_dict['columns'] = model_admin.get_columns(request)
else:
model_dict['columns'] = []
if 'is_tree' in dir(model_admin):
model_dict.update({'is_tree': True})
if app_label in app_dict:
app_dict[app_label]['models'].append(model_dict)
else:
app_dict[app_label] = {
'app': app_label,
'name': app_name,
'app_url': reverse('admin:app_list', kwargs={'app_label': app_label},
current_app=admin.site.name),
'has_module_perms': has_module_perms,
'models': [model_dict],
}
# Sort the apps alphabetically.
app_list = list(app_dict.values())
app_list.sort(key=lambda x: x['name'])
return app_list
def get_actions(request, model, model_admin):
actions = model_admin.get_action_choices(request)
template_actions = []
ct = ContentType.objects.get_for_model(model, not model._meta.proxy)
templates = Template.objects.filter(model=ct)
for template in templates:
template_actions.append({
'text': template.name,
'name': 'create_doc_' + template.name
})
if template_actions:
actions.append(('create_doc', 'Сделать документ', template_actions))
return actions
| [
"maxim.sorokin@gmail.com"
] | maxim.sorokin@gmail.com |
29dcc6918e6e732253f2ff0f81d976da35166127 | 744e72e6293c4cf93e1d97dc22c1d31851348056 | /resizePDF.py | a26bd81a5ec08659f044cac572a1f888ef3618db | [] | no_license | blueskycorner/resizePDF | b89fc58c380e4d795cf8dfa9ece4bf73220d38c4 | ddd479ad5e732141fe816a4ce1e188cb75e3dc42 | refs/heads/master | 2022-12-15T22:09:26.808343 | 2019-03-18T13:20:08 | 2019-03-18T13:20:08 | 173,715,753 | 0 | 0 | null | 2022-11-22T03:27:38 | 2019-03-04T09:36:43 | Dockerfile | UTF-8 | Python | false | false | 5,505 | py | import json
import os
import boto3
from fpdf import FPDF
from email.mime.text import MIMEText
from email.mime.application import MIMEApplication
from email.mime.multipart import MIMEMultipart
from botocore.exceptions import ClientError
bucketNameParamName = "bucketName"
tmpPathParamName = "tmpPath"
bucketName = os.getenv(bucketNameParamName)
s3 = boto3.client('s3')
def buildPDF(filename, imagesList):
try:
pdf = FPDF()
pdf.compress = False
print("Nb images to add: " + str(len(imagesList)))
for image in imagesList:
pdf.add_page()
pdf.image(image,0,0,210,297)
pdf.output(filename)
except Exception as e:
print(e)
raise e
def sendEmail(emailFrom, emailTo, downloadUrl):
print("address: " + emailTo)
print("downloadUrl: " + downloadUrl)
SENDER = emailFrom
RECIPIENT = emailTo
# The subject line for the email.
SUBJECT = "Your PDF file is ready !"
# The email body for recipients with non-HTML email clients.
BODY_TEXT = "Hi,\n\nHere is a link to download your file:\n" + downloadUrl + "\n\nHave a nice day."
# The HTML body of the email.
BODY_HTML = """\
<html>
<head></head>
<body>
Hi,<br><br>
Here is a link to download your document: <a href="
"""
BODY_HTML += downloadUrl
BODY_HTML += """\
">Click here</a>
<br><br>
Have a nice day.<br>
</body>
</html>
"""
# The character encoding for the email.
CHARSET = "utf-8"
# Create a new SES resource and specify a region.
client = boto3.client('ses')
# Create a multipart/mixed parent container.
msg = MIMEMultipart('mixed')
# Add subject, from and to lines.
msg['Subject'] = SUBJECT
msg['From'] = SENDER
msg['To'] = RECIPIENT
# Create a multipart/alternative child container.
msg_body = MIMEMultipart('alternative')
# Encode the text and HTML content and set the character encoding. This step is
# necessary if you're sending a message with characters outside the ASCII range.
textpart = MIMEText(BODY_TEXT.encode(CHARSET), 'plain', CHARSET)
htmlpart = MIMEText(BODY_HTML.encode(CHARSET), 'html', CHARSET)
# Add the text and HTML parts to the child container.
msg_body.attach(textpart)
msg_body.attach(htmlpart)
# Attach the multipart/alternative child container to the multipart/mixed
# parent container.
msg.attach(msg_body)
#print(msg)
try:
#Provide the contents of the email.
response = client.send_raw_email(
Source=SENDER,
Destinations=[
RECIPIENT
],
RawMessage={
'Data':msg.as_string(),
}
# ConfigurationSetName=CONFIGURATION_SET
)
# Display an error if something goes wrong.
except ClientError as e:
print(e.response['Error']['Message'])
else:
print("Email sent! Message ID:")
print(response['MessageId'])
def resizePDF(event, context):
response = None
try:
emailAddress = event['queryStringParameters']['emailAddress']
prefix = event['queryStringParameters']['prefix']
compression = event['queryStringParameters']['compression']
tmpPath = os.getenv(tmpPathParamName)
print("prefix: " + prefix)
print("compression: " + compression)
print("bucketName: " + bucketName)
# Download files and build
s3Ressource = boto3.resource('s3')
bucket = s3Ressource.Bucket(bucketName)
extensionsAllowed = ['.jpg', '.JPG', '.jpeg', '.JPEG', '.png', '.PNG']
imagesList = []
for object in bucket.objects.filter(Prefix=prefix):
print(object)
if (object.size > 0):
path, filename = os.path.split(object.key)
filenameshort, extension = os.path.splitext(filename)
print("path: " + path)
print("filename: " + filename)
print("filenameshort: " + filenameshort)
print("extension: " + extension)
if (extension in extensionsAllowed):
print("path: " + path)
print("filename: " + filename)
dest = tmpPath + filename
print("dest: " + dest)
bucket.download_file(object.key, dest)
imagesList.append(dest)
# Build the PDF
doc = tmpPath + prefix + ".pdf"
buildPDF(doc, imagesList)
destKey = prefix + "/document.pdf"
bucket.upload_file(doc, destKey)
signedUrlDownload = s3.generate_presigned_url(
ClientMethod='get_object',
Params={
'Bucket': bucketName,
'Key': destKey
},
ExpiresIn=3600
)
# Send the email
sendEmail(emailAddress, emailAddress, signedUrlDownload)
responseBody = {"signedUrlDownload": signedUrlDownload}
response = {
"statusCode": 200,
"body": json.dumps(responseBody),
"headers": {"Access-Control-Allow-Origin": "*"},
"isBase64Encoded": "false"
}
except Exception as e:
responseBody = {"error": str(e)}
response = {
"statusCode": 500,
"body": json.dumps(responseBody)
}
return response
| [
"ec2-user@ip-172-31-1-156.ec2.internal"
] | ec2-user@ip-172-31-1-156.ec2.internal |
f837b00ff86d2477efe671f6b6412d0ad0150621 | acb8e84e3b9c987fcab341f799f41d5a5ec4d587 | /langs/0/d1.py | 02a7c272c8bcf2a81b93b768bcfc238aa6155369 | [] | no_license | G4te-Keep3r/HowdyHackers | 46bfad63eafe5ac515da363e1c75fa6f4b9bca32 | fb6d391aaecb60ab5c4650d4ae2ddd599fd85db2 | refs/heads/master | 2020-08-01T12:08:10.782018 | 2016-11-13T20:45:50 | 2016-11-13T20:45:50 | 73,624,224 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 485 | py | import sys
def printFunction(lineRemaining):
if lineRemaining[0] == '"' and lineRemaining[-1] == '"':
if len(lineRemaining) > 2:
#data to print
lineRemaining = lineRemaining[1:-1]
print ' '.join(lineRemaining)
else:
print
def main(fileName):
with open(fileName) as f:
for line in f:
data = line.split()
if data[0] == 'D1':
printFunction(data[1:])
else:
print 'ERROR'
return
if __name__ == '__main__':
main(sys.argv[1]) | [
"juliettaylorswift@gmail.com"
] | juliettaylorswift@gmail.com |
25d08dc751e8d64a9112ba62276617681a6e5d78 | 42e0305c8cc9e20fee14d359ec3d466fb4608607 | /进程和线程/信号量.py | 96d8b44d2b620d8c9a64ba0f291d7ef73961073b | [] | no_license | wuxvsuizhong/Li_pro | 976159583927823464d4576efb59aaf86ef65e13 | 7facd87e67f767412917d9b8668746f1d87ec28f | refs/heads/master | 2023-08-08T23:13:08.226873 | 2023-07-22T10:09:25 | 2023-07-22T10:09:25 | 107,368,788 | 0 | 0 | null | 2017-10-18T06:50:33 | 2017-10-18T06:42:18 | null | UTF-8 | Python | false | false | 551 | py | import random
import time
from threading import Thread,Semaphore
# 信号量用于控制访问特定资源的线程数量,通常用于某些资源有明确访问数量限制额场景,简单说就是用于限流
# 停车场
# 设置车位资源数为5
sp = Semaphore(5)
def task(name):
sp.acquire()
print(f"{name}抢到了车位")
time.sleep(random.randint(3,5))
sp.release()
print(f"{name}开走了")
if "__main__" == __name__:
for i in range(10):
t = Thread(target=task,args=(f"宝马{i}",))
t.start() | [
"devuser01@123.com"
] | devuser01@123.com |
8c8a9ad367a45afbe6c7fd1fd503d6ef9dc05db6 | 901562de637c44a0f0c43aeebe56f5ebd949571e | /analysis/repo-statistics/project_counters_jarsize_tab.py | 79e19b8a948f5b16eeae9fe84647de8d5cf8bf77 | [] | no_license | istlab/evol_security_publication_2012 | f41608dd7d054aa890fa6d011dd3a44407b3a598 | daa64b60641a82fb5ed9a5b4e8f08c75f700138c | refs/heads/master | 2020-12-11T09:31:03.511279 | 2013-11-10T23:02:44 | 2013-11-10T23:02:44 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,744 | py | # Extracts the bug counters data from the JSON representation
# and outputs them in CSV format so that it can be read by R.
import ijson
import json
import sys
import csv
bug_types = [
'SECURITY_HIGH',
'SECURITY_LOW',
'STYLE',
'CORRECTNESS',
'BAD_PRACTICE',
'MT_CORRECTNESS',
'I18N',
'PERFORMANCE',
'EXPERIMENTAL',
]
with open("data/project_counters_jarsize.csv", "w") as csv_output:
csvwriter = csv.writer(csv_output)
project_counts = {}
project_key = ""
row = ['project', 'version', 'jarsize']
for bug_type in bug_types:
row.append(bug_type)
csvwriter.writerow(row)
with open("data/project_counters.json", "r") as json_file:
json_input = json.load(json_file)
for project, data in json_input.iteritems():
for version in data['versions']:
meta_data = version['JarMetadata']
row = [project, meta_data['version_order'],
meta_data['jar_size']]
counters = version['Counters']
counters = version['Counters']
if 'MALICIOUS_CODE' in counters:
malicious_code = counters.pop('MALICIOUS_CODE')
if 'SECURITY_LOW' in counters:
counters['SECURITY_LOW'] += malicious_code
else:
counters['SECURITY_LOW'] = malicious_code
security_low = 0
for bug_type in bug_types:
if bug_type in counters:
row.append(counters[bug_type])
else:
row.append('NA')
csvwriter.writerow(row)
| [
"louridas@aueb.gr"
] | louridas@aueb.gr |
880bc21c593bf7e8364d6d0ac67964c6954c756b | 754a22e0156af095d224075cfa2a1a460014756f | /mysite/settings.py | 70dd40e6c9fc7427dc6d17bc4823dd416a4d8cf6 | [] | no_license | LucieGal/my-own-blog | 3a7fa4a7bcaf80db489777325bb36df2cee5bfd8 | f0f07f24e118e6e9e8c97474526c8ea5eb843f60 | refs/heads/master | 2023-06-15T00:52:30.462831 | 2021-07-10T14:03:59 | 2021-07-10T14:03:59 | 379,363,073 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,205 | py | """
Django settings for mysite project.
Generated by 'django-admin startproject' using Django 2.2.24.
For more information on this file, see
https://docs.djangoproject.com/en/2.2/topics/settings/
For the full list of settings and their values, see
https://docs.djangoproject.com/en/2.2/ref/settings/
"""
import os
# Build paths inside the project like this: os.path.join(BASE_DIR, ...)
BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
# Quick-start development settings - unsuitable for production
# See https://docs.djangoproject.com/en/2.2/howto/deployment/checklist/
# SECURITY WARNING: keep the secret key used in production secret!
SECRET_KEY = '%ufdj@-rtsl93rv1o8xvev=+b(cvm$5tbaayzt7$*(2dsr-#ds'
# SECURITY WARNING: don't run with debug turned on in production!
DEBUG = True
ALLOWED_HOSTS = ['127.0.0.1', '.pythonanywhere.com']
# Application definition
INSTALLED_APPS = [
'django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'blog.apps.BlogConfig',
]
MIDDLEWARE = [
'django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
]
ROOT_URLCONF = 'mysite.urls'
TEMPLATES = [
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.template.context_processors.debug',
'django.template.context_processors.request',
'django.contrib.auth.context_processors.auth',
'django.contrib.messages.context_processors.messages',
],
},
},
]
WSGI_APPLICATION = 'mysite.wsgi.application'
# Database
# https://docs.djangoproject.com/en/2.2/ref/settings/#databases
DATABASES = {
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': os.path.join(BASE_DIR, 'db.sqlite3'),
}
}
# Password validation
# https://docs.djangoproject.com/en/2.2/ref/settings/#auth-password-validators
AUTH_PASSWORD_VALIDATORS = [
{
'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator',
},
{
'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator',
},
]
# Internationalization
# https://docs.djangoproject.com/en/2.2/topics/i18n/
LANGUAGE_CODE = 'fr-fr'
TIME_ZONE = 'Asia/Tokyo'
USE_I18N = True
USE_L10N = True
USE_TZ = True
# Static files (CSS, JavaScript, Images)
# https://docs.djangoproject.com/en/2.2/howto/static-files/
STATIC_URL = '/static/'
STATIC_ROOT = os.path.join(BASE_DIR, 'static')
| [
"galea.lucie@gmail.com"
] | galea.lucie@gmail.com |
07ebc735e3d7c3b9475a58511601718c8e2eb1b3 | 3cd0623135f5005eeec232f508f7f3f523e156f3 | /fiddles/_08/test_fiddle.py | 5cfdac2ab8ec333514e8bcb737073bb706cd94ec | [] | no_license | tompascall/adventofcode2019 | 4f287ec7cbe0f97caed68fb88ed54c552766f56c | 8b2eebf902e833029aecd508343b3d8b3a086460 | refs/heads/master | 2020-09-28T08:25:37.828312 | 2020-03-22T09:19:15 | 2020-03-22T09:19:15 | 226,733,027 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,487 | py | import unittest
from fiddles._08.fiddle import (
get_layers,
count_digit_in_layer,
merge_layers,
)
class TestFiddle(unittest.TestCase):
def test_get_layers(self):
digits = '123456789012'
layers = get_layers(digits, width=3, height=2)
self.assertEqual(
layers,
[
[
['1','2','3'],
['4','5','6']
],
[
['7','8','9'],
['0','1','2']
]
]
)
def test_count_digit_in_layer(self):
digits = '121456789012'
layers = get_layers(digits, width=3, height=2)
self.assertEqual(
count_digit_in_layer(layers[0], '1'),
2
)
layers = [
[
['0','2'],
['2','2'],
['2','2']
],
[
['1','1'],
['2','2'],
['2','2']
],
[
['2','2'],
['1','2'],
['1','0']
],
[
['0','0'],
['0','0'],
['2','2']
],
]
self.assertEqual(
merge_layers(layers, 3),
[
['0','1'],
['1','0'],
['1','0']
]
)
if __name__ == '__main__':
unittest.main()
| [
"tamas.gulacsy.toth@lensa.com"
] | tamas.gulacsy.toth@lensa.com |
04d46f70d2543594d36fc9d340ad9c2da9f9cd7b | 7eb8bf846dc7021751019debf91925139203bed2 | /Django_Clases/tercer_proyecto/populate_modelos_aplicacion.py | 348b1e00929e50d9b01698e636df06708a4c9001 | [] | no_license | rpparada/python-and-django-full-stack-web-developer-bootcamp | 5c384dc1c19557097c893cf6149c1831984b1946 | 7b91f16cfb49d7de71901857b4e4c8f447db5e6f | refs/heads/master | 2021-09-08T22:40:44.737431 | 2018-03-12T15:12:06 | 2018-03-12T15:12:06 | 116,153,519 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 695 | py | import os
os.environ.setdefault('DJANGO_SETTINGS_MODULE','tercer_proyecto.settings')
import django
django.setup()
import random
from modelos_aplicacion.models import Usuarios
from faker import Faker
generaFake = Faker()
def popular(N=10):
for entrada in range(N):
nombre_falso = generaFake.first_name()
apellido_falso = generaFake.last_name()
email_falso = generaFake.email()
# email_falso = generaFake.email(*args, **kwargs)
usuario = Usuarios.objects.get_or_create(nombre=nombre_falso,apellido=apellido_falso,email=email_falso)[0]
if __name__ == '__main__':
print('Cargando tabla(s)... ')
popular(20)
print('Rabla(s) cargada(s)!')
| [
"rpparada@gmail.com"
] | rpparada@gmail.com |
fe01b307a0814fd473a553ad5bfd3a7ad7f22547 | 245a3f8cea6f232bf3142706c11188b51eb21774 | /python/hetu/onnx/onnx_opset/Where.py | 6da3b659d9f9858f398695ae791903a6f8c2c8b5 | [
"Apache-2.0"
] | permissive | initzhang/Hetu | 5bfcb07e62962fbc83def14148f8367fab02625a | 447111a358e4dc6df5db9c216bdb3590fff05f84 | refs/heads/main | 2023-06-20T18:37:21.760083 | 2021-07-27T04:37:48 | 2021-07-27T04:37:48 | 389,848,768 | 0 | 0 | Apache-2.0 | 2021-07-27T04:32:57 | 2021-07-27T04:32:57 | null | UTF-8 | Python | false | false | 610 | py | from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
import numpy as np
from onnx import onnx_pb
from hetu.onnx import constants, util, graph
from hetu.onnx.handler import hetu_op
from hetu.onnx.onnx_opset import general
@hetu_op(["WhereOp"], onnx_op=["Where"])
class Where():
@classmethod
def version_1(cls, ctx, node, **kwargs):
assert False, "This version of the operator has been available since version 9 of the default ONNX operator set"
pass
@classmethod
def version_9(cls, ctx, node, **kwargs):
pass
| [
"swordonline@foxmail.com"
] | swordonline@foxmail.com |
2ca77983524514c47a936a1f296297e5ba1c4456 | 7b1b4ed8bd4c887362b367625a833c28aa919dd8 | /wpaudit/providers/aliyun/resources/ram/policies.py | 09ac9427cfcba323da87129ef7e60ece906a9935 | [] | no_license | wperic/wpaudit | 6bbd557c803ce9bceb764c1451daeb5e440a3d9c | ed69c1eabcf85e80ed8fe5397d2d369fd3ff35d8 | refs/heads/main | 2023-07-16T21:36:57.528548 | 2021-09-03T10:35:43 | 2021-09-03T10:35:43 | 402,716,870 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,794 | py | from wpaudit.providers.aliyun.resources.base import AliyunResources
from wpaudit.providers.aliyun.facade.base import AliyunFacade
import json
class Policies(AliyunResources):
def __init__(self, facade: AliyunFacade):
super().__init__(facade)
async def fetch_all(self):
for raw_policy in await self.facade.ram.get_policies():
id, policy = await self._parse_policy(raw_policy)
if id:
self[id] = policy
async def _parse_policy(self, raw_policy):
"""
Only processing policies with an
:param raw_policy:
:return:
"""
if raw_policy.get('AttachmentCount') > 0:
policy_dict = {}
policy_dict['id'] = policy_dict['name'] = raw_policy.get('PolicyName')
policy_dict['description'] = raw_policy.get('Description')
policy_dict['create_date'] = raw_policy.get('CreateDate')
policy_dict['update_date'] = raw_policy.get('UpdateDate')
policy_dict['attachment_count'] = raw_policy.get('AttachmentCount')
policy_dict['type'] = raw_policy.get('PolicyType')
policy_dict['default_version'] = raw_policy.get('DefaultVersion')
policy_version = await self.facade.ram.get_policy_version(policy_dict['name'],
policy_dict['type'],
policy_dict['default_version'])
policy_version['PolicyDocument'] = json.loads(policy_version['PolicyDocument'])
# policy_dict['policy_document'] = policy_version['PolicyDocument']
policy_dict['policy_document'] = policy_version
policy_entities = await self.facade.ram.get_policy_entities(policy_dict['name'],
policy_dict['type'])
policy_dict['entities'] = {}
if policy_entities['Users']['User']:
policy_dict['entities']['users'] = []
for user in policy_entities['Users']['User']:
policy_dict['entities']['users'].append(user['UserName'])
if policy_entities['Groups']['Group']:
policy_dict['entities']['groups'] = []
for group in policy_entities['Groups']['Group']:
policy_dict['entities']['groups'].append(group['GroupName'])
if policy_entities['Roles']['Role']:
policy_dict['entities']['roles'] = []
for role in policy_entities['Roles']['Role']:
policy_dict['entities']['roles'].append(role['RoleName'])
return policy_dict['id'], policy_dict
else:
return None, None
| [
"90035639+wperic@users.noreply.github.com"
] | 90035639+wperic@users.noreply.github.com |
752a144606ef7beeca941dda8a76f1980182e009 | 3a6fb0ef104d07491cbb56d95f7a6140dff29eaa | /2b.py | a97428ac04428c13b7ffa2cfada1e457d2e9693e | [] | no_license | audiodude/advent2019 | 5903e38f87d095c8b2bd30be7a7a18bbc8b25310 | 4c725a732c5691df60f99d4eda562b0cf81020fc | refs/heads/master | 2020-12-01T02:03:07.433484 | 2019-12-30T18:55:22 | 2019-12-30T18:55:22 | 230,537,406 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 1,953 | py | # First attempt: reverse the program and find the combination of initial values
# that produce the output. Unfortunately, we of course end up with output like this:
# (+ (+ -2 (+ 4 (+ 2 (+ 1 (+ (+ 3 (* (+ 5 (+ (+ 2 (* (+ (+ 3 (* 5 (+ 2 (* (+ 2
# (+ (* 2 (* (* (+ (* (+ 1 (* 4 (+ (+ 1 (* (+ 2 (* 4 -1)) 2)) 3))) 4) 2) 2) 5
# )) 1)) 3)))) 5) 4)) 5)) 4)) 4))))) 4)
#
# Which is a linear equation with two unknowns (-1 and -2 in this case) which
# of course can't be solved.
#
# import fileinput
# codes = [int(c) for c in fileinput.input()[0].split(',')]
# codes[1] = -1
# codes[2] = -2
# stop_pos = 0
# for pos in range(len(codes) - 1, 0, -1):
# if codes[pos] == 99:
# stop_pos = pos
# break
# assert stop_pos, 'stop_pos not found'
# def value_of(idx, stop_pos):
# for pos in range(stop_pos, 0, -4):
# opcode, p1, p2, out = codes[pos-4:pos]
# if out != idx:
# continue
# if opcode == 1:
# return '(+ %s %s)' % (value_of(p1, pos-4), value_of(p2, pos-4))
# elif opcode == 2:
# return '(* %s %s)' % (value_of(p1, pos-4), value_of(p2, pos-4))
# else:
# return str(codes[idx])
# print(value_of(0, stop_pos))
# Second attempt, brute force guessing:
import fileinput
def process(codes):
for pos in range(0, len(codes), 4):
if codes[pos] == 99:
break
op1 = codes[codes[pos+1]]
op2 = codes[codes[pos+2]]
output_pos = codes[pos + 3]
if codes[pos] == 1:
codes[output_pos] = op1 + op2
elif codes[pos] == 2:
codes[output_pos] = op1 * op2
codes = [int(c) for c in fileinput.input()[0].split(',')]
def attempt(tries):
for noun in range(tries):
for verb in range(tries):
new_codes = codes[:]
new_codes[1] = noun
new_codes[2] = verb
process(new_codes)
if new_codes[0] == 19690720:
return (noun, verb)
else:
assert False, 'Could not find answer'
noun, verb = attempt(100)
print(noun * 100 + verb)
| [
"audiodude@gmail.com"
] | audiodude@gmail.com |
743cc0818768c373bc08f9acf81e567aacb3a69b | d528d21d32a2a7f299e8365d0a935b8718f9c07f | /cogs/utils/checks.py | 7f0962fe5b0e94c665e2849f9eb198a293c99c7d | [] | no_license | sizumita/Aegis | 53b3f3db4d88b8ffdbc0d44781f55251081a32fc | 2c9684695a32481583fd214fa63deaddea3d5ebc | refs/heads/master | 2020-09-11T00:05:48.629459 | 2020-06-23T14:04:41 | 2020-06-23T14:04:41 | 221,874,644 | 6 | 4 | null | 2019-12-10T10:58:34 | 2019-11-15T08:04:23 | Python | UTF-8 | Python | false | false | 2,758 | py | from .database import CommandPermission
from discord.ext.commands import check
import discord
async def check_command_permission(context):
"""
権限周りについて:
DMの場合確実に有効
CommandPermissionがなければそもそも有効化されていない
作成されていて、かつroles、users、permissionsが空であれば誰でも使える
:param context: commands.Context
:return: bool
"""
# DMの場合
if not context.guild:
return True
# manage系、ヘルプコマンドだった場合
if context.command.name == 'help':
return True
elif context.cog:
if context.cog.qualified_name == 'Manage':
return True
p: CommandPermission = await CommandPermission.query.where(CommandPermission.id == context.guild.id) \
.where(CommandPermission.name == context.bot.get_command_full_name(context.command)).gino.first()
# ない場合
if not p:
if getattr(context.cog, 'already_on', False):
p = await CommandPermission.create(id=context.guild.id,
name=context.bot.get_command_full_name(context.command))
else:
return False
if context.author.guild_permissions.administrator:
return True
# 制限なしの場合
if not p.roles and not p.users:
return True
checks = []
if p.roles:
is_id_in = any(True for i in context.author.roles if str(i.id) in p.roles)
checks.append(is_id_in)
if p.users:
checks.append(True if str(context.author.id) in p.users else False)
return any(checks)
def admin_only():
def predicate(ctx):
permissions: discord.Permissions = ctx.author.guild_permissions
if not permissions.administrator:
return False
return True
return check(predicate)
def safety():
"""CommandPermissionがあってかつ何も設定されていないときにadminしか実行できないようにする"""
async def predicate(ctx):
p: CommandPermission = await CommandPermission.query.where(CommandPermission.id == ctx.guild.id) \
.where(CommandPermission.name == ctx.bot.get_command_full_name(ctx.command)).gino.first()
if not p:
return False
if not p.users and not p.roles:
permissions: discord.Permissions = ctx.author.guild_permissions
if not permissions.administrator:
return False
return True
return check(predicate)
def prefix_in(prefixes):
async def predicate(ctx):
if ctx.prefix not in prefixes:
return False
return True
return check(predicate)
| [
"sumito@izumita.com"
] | sumito@izumita.com |
3d12d6287b41ff9445a633794470f5b2b34fd4ee | c45306676df2fe733007ff12acd88a8b7d700c42 | /DrawLossCurve.py | 3f600ae3be6c244da132e775e4fcbcf478a1c218 | [] | no_license | huanhsu/Caffe_Tool | 83863ea351da4d3006f2ad4961b91e5b0c758239 | bc9d7205f26ed3e071b1f2d3fac4ab14d64b82e4 | refs/heads/master | 2021-01-16T17:55:23.472297 | 2017-08-16T07:20:59 | 2017-08-16T07:20:59 | 100,025,467 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,751 | py | import numpy as np
import re
import matplotlib.patches as mpatches
from argparse import ArgumentParser
from matplotlib import pylab as plt
import pdb
#Draw loss curves with many output in a file
#This example has 3 loss curves
#an example on log
#I0728 11:05:30.162691 1057 solver.cpp:219] Iteration 20 (0.0501322 iter/s, 398.945s/20 iters), loss = 29.1468
#I0728 11:05:30.162755 1057 solver.cpp:238] Train net output #0: prob = 7.30188 (* 1 = 7.30188 loss)
#I0728 11:05:30.162770 1057 solver.cpp:238] Train net output #1: prob_2c_4f = 7.30041 (* 1 = 7.30041 loss)
#I0728 11:05:30.162781 1057 solver.cpp:238] Train net output #2: prob_2c_5c = 7.29855 (* 1 = 7.29855 loss)
#I0728 11:05:30.162792 1057 solver.cpp:238] Train net output #3: prob_3d_5c = 7.29898 (* 1 = 7.29898 loss)
#explaining for pattern
#([+-]?(\d+(\.\d*)?|\.\d+)([eE][+-]?\d+)?) is matching number with point or number with e
#.*\n.* means no matter string after, new line, no matter string before
def main(files):
plt.style.use('ggplot')
fig, ax1 = plt.subplots()
ax1.set_ylim([0.002,8])
ax1.set_xlabel('Iteration')
ax1.set_ylabel('Training Loss')
temp=[]
temp.append(files.files1)
for i, log_file in enumerate(temp):
loss_iterations, losses, losses2, losses3, losses4 = parse_log(log_file)
disp_results(fig, ax1, loss_iterations, losses, losses2, losses3, losses4, color_ind=i)
patch1 = mpatches.Patch(color=colors[0], label=' loss')
patch2 = mpatches.Patch(color=colors[1], label='2c_4f loss')
patch3 = mpatches.Patch(color=colors[2], label='2c_5c loss')
patch4 = mpatches.Patch(color=colors[3], label='3d_5c loss')
plt.legend(handles=[patch1, patch2, patch3, patch4])
plt.show()
def parse_log(log_file):
with open(log_file, 'r') as log_file:
log = log_file.read()
loss_pattern = r"Iteration (\d+) \(.*\), loss = .*\n.* Train net output #0: prob = ([+-]?(\d+(\.\d*)?|\.\d+)([eE][+-]?\d+)?) .*\n.* Train net output #1: prob_2c_4f = ([+-]?(\d+(\.\d*)?|\.\d+)([eE][+-]?\d+)?) .*\n.* Train net output #2: prob_2c_5c = ([+-]?(\d+(\.\d*)?|\.\d+)([eE][+-]?\d+)?) .*\n.* Train net output #3: prob_3d_5c = ([+-]?(\d+(\.\d*)?|\.\d+)([eE][+-]?\d+)?)"
losses = []
losses2 = []
losses3 = []
losses4 = []
loss_iterations = []
for r in re.findall(loss_pattern, log):
loss_iterations.append(int(r[0]))
losses.append(float(r[1]))
losses2.append(float(r[5]))
losses3.append(float(r[9]))
losses4.append(float(r[13]))
loss_iterations = np.array(loss_iterations)
losses = np.array(losses)
losses2 = np.array(losses2)
losses3 = np.array(losses3)
losses4 = np.array(losses4)
return loss_iterations, losses, losses2, losses3, losses4
def disp_results(fig, ax1, loss_iterations, losses, losses2, losses3, losses4, color_ind=0):
modula = len(plt.rcParams['axes.color_cycle'])
color1=plt.rcParams['axes.color_cycle'][(color_ind * 4 + 0) % modula]
ax1.plot(loss_iterations, losses,color=color1)
color2=plt.rcParams['axes.color_cycle'][(color_ind * 4 + 1) % modula]
ax1.plot(loss_iterations, losses2,color=color2)
color3=plt.rcParams['axes.color_cycle'][(color_ind * 4 + 2) % modula]
ax1.plot(loss_iterations, losses3,color=color3)
color4=plt.rcParams['axes.color_cycle'][(color_ind * 4 + 3) % modula]
ax1.plot(loss_iterations, losses4,color=color4)
colors.append(color1)
colors.append(color2)
colors.append(color3)
colors.append(color4)
if __name__ == '__main__':
global colors
colors=[]
parser = ArgumentParser(description="Draw loss curve")
parser.add_argument('files1')
args = parser.parse_args()
main(args)
| [
"noreply@github.com"
] | noreply@github.com |
9dab3cde69c85a1c6fe735d132871bf0bf28607e | c9f0975de0e1bfe9043ef04d43df00e6c7fcbe56 | /src/bandits/__init__.py | b318a2379fd7830e1500ce648db25ce5d8531d14 | [] | no_license | kbantoec/bandits | dbaf89d4c29a14a97e2d39539410874d3ff582bc | 4585d0c1b60c19cb2511cabfb1ce211499e72ebe | refs/heads/master | 2023-01-19T04:41:32.855694 | 2020-11-10T18:11:25 | 2020-11-10T18:11:25 | 306,442,333 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 82 | py | from .agent import Bandit
from .one_armed_bandit import OneArmedBandit, Experiment | [
"y.banto@outlook.com"
] | y.banto@outlook.com |
daf9960e7287238bf5dffe7767564296ba29dcd1 | f46fee7ac51bc459f6309d5dafed1a277cc52772 | /demo16.py | 2990ce342a9f5eda4c54f8c5d24ecbf8d9ebe394 | [] | no_license | wu840407/PYKT-Python-for-Keras-and-TensorFlow | d7e0637640b669246bc7f94e32ee32af1a23d375 | 90c5305f7167def16202626c99242978a829d2e0 | refs/heads/master | 2023-05-01T05:47:47.772704 | 2021-05-14T09:02:17 | 2021-05-14T09:02:17 | 366,560,311 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 734 | py | import matplotlib.pyplot as plt
from sklearn import datasets
iris = datasets.load_iris()
print(dir(iris))
labels = iris.feature_names
print(labels)
X = iris.data
species = iris.target
counter = 1
for i in range(0, 4):
for j in range(i + 1, 4):
plt.figure(counter, figsize=(12, 9))
counter += 1
xData = X[:, i]
yData = X[:, j]
x_min, x_max = xData.min() - 0.5, xData.max() + 0.5
y_min, y_max = yData.min() - 0.5, yData.max() + 0.5
plt.scatter(xData, yData, c=species, cmap=plt.cm.Paired)
plt.xlabel(labels[i])
plt.ylabel(labels[j])
plt.xlim(x_min, x_max)
plt.ylim(y_min, y_max)
plt.xticks([])
plt.yticks([])
plt.show() | [
"wu@uuu.com.tw"
] | wu@uuu.com.tw |
a92e09e86e4b6a1367be78bdfb2c56bb263cd6dc | ed702ade0903ae74709da52fca3b3279c45877c7 | /Processors/translation.py | 113548142dae0d1829316206661dc96f7a83fbc6 | [] | no_license | aaskorohodov/pythonProject3 | 1611f0846a4f9af3ee12a5b7b62b56beb5d22791 | 5257c41770a61d0351b65bfe25bca0a2c52ac3c6 | refs/heads/master | 2023-07-30T22:51:45.839766 | 2021-08-27T20:52:49 | 2021-08-27T20:55:01 | 393,996,864 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,023 | py | import requests
def ru_eng(message):
word = message.text
URL_AUTH = 'https://developers.lingvolive.com/api/v1.1/authenticate'
URL_TRANSLATE = 'https://developers.lingvolive.com/api/v1/Minicard'
KEY = 'ODRlNGU0NDctOWU2ZS00ZTcxLTk5ZWYtNjI2ZTA0MzYwOGU3OmQwN2E5YjVhMDg2MjQwYTI5ZDU3NjA1Y2NiODI3ZjRj'
headers_auth = {'Authorization': 'Basic ' + KEY}
auth = requests.post(URL_AUTH, headers=headers_auth)
if auth.status_code == 200:
token = auth.text
headers_translate = {'Authorization': 'Bearer ' + token}
params = {'text': word,
'srcLang': 1049,
'dstLang': 1033}
r = requests.get(URL_TRANSLATE, headers=headers_translate, params=params)
r = r.json()
try:
return r['Translation']['Translation']
except:
return 'Перевод не найден'
else:
return 'Сервис перевода сейчас недоступен. Попробуйте позднее'
def eng_ru(message):
word = message.text
URL_AUTH = 'https://developers.lingvolive.com/api/v1.1/authenticate'
URL_TRANSLATE = 'https://developers.lingvolive.com/api/v1/Minicard'
KEY = 'ODRlNGU0NDctOWU2ZS00ZTcxLTk5ZWYtNjI2ZTA0MzYwOGU3OmQwN2E5YjVhMDg2MjQwYTI5ZDU3NjA1Y2NiODI3ZjRj'
headers_auth = {'Authorization': 'Basic ' + KEY}
auth = requests.post(URL_AUTH, headers=headers_auth)
if auth.status_code == 200:
token = auth.text
headers_translate = {'Authorization': 'Bearer ' + token}
params = {'text': word,
'srcLang': 1033,
'dstLang': 1049}
r = requests.get(URL_TRANSLATE, headers=headers_translate, params=params)
r = r.json()
try:
return r['Translation']['Translation']
except:
return 'Перевод не найден'
else:
return 'Сервис перевода сейчас недоступен. Попробуйте позднее' | [
"aaskorohodov@gmail.com"
] | aaskorohodov@gmail.com |
ae2eade74f9f078d1840f1f5df750227c8959659 | ce6e91fb9a5a9049d817d020ca0018b7f4008b9b | /runtests.py | ef35cd877b6d81a7ad6d506365c6d7dfbe0e8cb7 | [] | no_license | ccnmtl/django-pagetimer | b98536273b38c64f10d6832b7b74833099e68436 | 2844b3c702df2952deffdf6cd75c9e47e6f35284 | refs/heads/master | 2021-01-09T20:53:18.627185 | 2017-08-30T19:32:23 | 2017-08-30T19:32:23 | 58,394,973 | 0 | 0 | null | 2017-08-30T19:32:23 | 2016-05-09T17:25:37 | Python | UTF-8 | Python | false | false | 2,149 | py | """ run tests for pagetimer
$ virtualenv ve
$ ./ve/bin/pip install Django==1.8
$ ./ve/bin/pip install .
$ ./ve/bin/python runtests.py
"""
import django
from django.conf import settings
from django.core.management import call_command
def main():
# Dynamically configure the Django settings with the minimum necessary to
# get Django running tests
settings.configure(
MIDDLEWARE_CLASSES=(
'django.contrib.sessions.middleware.SessionMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
),
INSTALLED_APPS=(
'django.contrib.auth',
'django.contrib.sessions',
'django.contrib.contenttypes',
'pagetimer',
),
TEST_RUNNER='django.test.runner.DiscoverRunner',
TEMPLATES=[
{
'BACKEND': 'django.template.backends.django.DjangoTemplates',
'DIRS': [],
'APP_DIRS': True,
'OPTIONS': {
'context_processors': [
'django.contrib.auth.context_processors.auth',
'django.template.context_processors.debug',
'django.template.context_processors.i18n',
'django.template.context_processors.media',
'django.template.context_processors.static',
'django.template.context_processors.tz',
'django.contrib.messages.context_processors.messages',
],
},
},
],
COVERAGE_EXCLUDES_FOLDERS=['migrations'],
ROOT_URLCONF='pagetimer.urls',
# Django replaces this, but it still wants it. *shrugs*
DATABASES={
'default': {
'ENGINE': 'django.db.backends.sqlite3',
'NAME': ':memory:',
'HOST': '',
'PORT': '',
'USER': '',
'PASSWORD': '',
}
},
)
django.setup()
# Fire off the tests
call_command('test')
if __name__ == '__main__':
main()
| [
"anders@columbia.edu"
] | anders@columbia.edu |
8a20be67dc8ff938517418a5e77c313ae3bb79ea | 318289d6d338b5aa7d87df18095839fa679ec588 | /ksusta/apps.py | 75b5219e7f210a93577efe0f443c56064e3e132d | [
"MIT"
] | permissive | ayubaezekiel/eExams | b491f36649d24ad6d87bc30017d09117c52ad926 | 5c9920cf55b9cb258c7c732bffceb10e5e9188b5 | refs/heads/main | 2023-09-05T14:05:11.134918 | 2021-11-08T11:41:17 | 2021-11-08T11:41:17 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 144 | py | from django.apps import AppConfig
class KsustaConfig(AppConfig):
default_auto_field = 'django.db.models.BigAutoField'
name = 'ksusta'
| [
"ezekielayoba@yahoo.com"
] | ezekielayoba@yahoo.com |
9f350befb965c94227bb57cfedbbedd959044200 | 6567a6d0b648300cc5c3fa264a925602f61ab8c4 | /guvi5.py | e5e6aefb6421834b1423ab36aa25c4ca25a15943 | [] | no_license | AnanthiD/codekata | 6ee948ca2aea9a052a1b4604e4fc28fb91b18cda | 533e2d0b9b3ca14c37eac936a927d9933eb35374 | refs/heads/master | 2020-05-23T01:05:28.676110 | 2019-07-20T09:45:59 | 2019-07-20T09:45:59 | 186,580,980 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 68 | py | a=input()
if(a.isalpha()):
print("Alphabet")
else:
print("no") | [
"noreply@github.com"
] | noreply@github.com |
0c124ba03429697b2bab6c819eb50dc44460c5dc | 84846fa2c35819f386a7b6f55f32227ff9d34838 | /work/random_likes.py | bd17d65524fbfdc4cba7a3c7253f6c373a27d032 | [
"Apache-2.0"
] | permissive | deniskolokol/universalrec | 43efcfc0c3b9d44e5e4a16bd27b9710228fa1085 | 55169f718777bb557d90c709c9a505a0a36f5478 | refs/heads/master | 2023-01-12T12:32:44.590950 | 2016-03-07T10:19:36 | 2016-03-07T10:19:36 | 51,755,830 | 0 | 0 | null | 2022-12-26T20:01:17 | 2016-02-15T13:03:23 | Scala | UTF-8 | Python | false | false | 791 | py | # run this first in flickthru virtualenv!
import os
import random
from core.models import Like
from django.db.models import Count
filename = 'likes.csv'
fobj = open(filename, 'w+')
fobj.write('entity_id,event,target_entity_id,event_time\n')
img_index = range(499) # indexes of images to pick from
registered = {}
for rec in Like.objects.order_by('image', 'user', 'liked', 'created_at'):
try:
image = registered[str(rec.image.id)]
except KeyError:
image = img_index.pop(random.randrange(len(img_index)))
registered[str(rec.image.id)] = image
event = 'like' if rec.liked else 'dislike'
fobj.write('%s,%s,%s,%s\n' % (
rec.user.id, event, image, rec.created_at.isoformat()
))
fobj.close()
print 'Done: %s' % os.path.abspath(filename)
| [
"dkolokol@gmail.com"
] | dkolokol@gmail.com |
94ec5975940892096bc5b805de5af3e9c66312a3 | 6b8960551ee4be37c46f6c5f28257845fcb871ed | /task1.py | 2105ae960977b9acf3bde10337df6d46c5ad633f | [] | no_license | htrueman/db2_limited_test | 10e9e574fe52b2346c33f4485f8b1dec00c30ac8 | 489379a952ad5c1ecb5123e9e3d41ec28206dc01 | refs/heads/master | 2022-12-09T06:32:27.709446 | 2017-06-12T01:40:08 | 2017-06-12T01:40:08 | 93,772,542 | 0 | 0 | null | 2022-11-22T01:46:27 | 2017-06-08T16:56:17 | Python | UTF-8 | Python | false | false | 649 | py | test_num1 = 1
test_num2 = 10
test_num3 = 2
def handle_numbers(number1, number2, number3):
count_div_numbers = 0
div_numbers_list = []
for number in range(number1, number2 + 1):
if number % number3 == 0:
count_div_numbers += 1
div_numbers_list.append(str(number))
if div_numbers_list:
return "Result:\n{}, because {} are divisible by {}".\
format(count_div_numbers, ', '.join(div_numbers_list), number3)
else:
return "Result:\nThere are no divisible numbers by {} in given range".\
format(number3)
print (handle_numbers(test_num1, test_num2, test_num3))
| [
"vege1wgw@gmail.com"
] | vege1wgw@gmail.com |
15203c190344a87445df55a4a7f5f3c15e73d4b5 | 03da0660db7c833e476dd4cba44ba2c300d3fb58 | /.PyCharmCE2017.3/system/python_stubs/-762174762/_hashlib.py | eb91b77cf05268f82a8144bddd9ef684816d9226 | [] | no_license | Shreyaskulkarni98/MQTTBroker-Beta | c9717d3e9d559f9f7d6eb46738dac68f8c4f346c | 22a15cac4755adf522d76e15129c0ef4b2da797d | refs/heads/master | 2022-11-01T20:58:40.690542 | 2019-01-20T08:51:26 | 2019-01-20T08:51:26 | 161,118,749 | 0 | 1 | null | 2022-10-22T14:32:03 | 2018-12-10T04:53:34 | Python | UTF-8 | Python | false | false | 1,710 | py | # encoding: utf-8
# module _hashlib
# from (pre-generated)
# by generator 1.145
# no doc
# no imports
# functions
def new(*args, **kwargs): # real signature unknown
"""
Return a new hash object using the named algorithm.
An optional string argument may be provided and will be
automatically hashed.
The MD5 and SHA1 algorithms are always supported.
"""
pass
def openssl_md5(*args, **kwargs): # real signature unknown
""" Returns a md5 hash object; optionally initialized with a string """
pass
def openssl_sha1(*args, **kwargs): # real signature unknown
""" Returns a sha1 hash object; optionally initialized with a string """
pass
def openssl_sha224(*args, **kwargs): # real signature unknown
""" Returns a sha224 hash object; optionally initialized with a string """
pass
def openssl_sha256(*args, **kwargs): # real signature unknown
""" Returns a sha256 hash object; optionally initialized with a string """
pass
def openssl_sha384(*args, **kwargs): # real signature unknown
""" Returns a sha384 hash object; optionally initialized with a string """
pass
def openssl_sha512(*args, **kwargs): # real signature unknown
""" Returns a sha512 hash object; optionally initialized with a string """
pass
def pbkdf2_hmac(hash_name, password, salt, iterations, dklen=None): # real signature unknown; restored from __doc__
"""
pbkdf2_hmac(hash_name, password, salt, iterations, dklen=None) -> key
Password based key derivation function 2 (PKCS #5 v2.0) with HMAC as
pseudorandom function.
"""
pass
# no classes
# variables with complex values
openssl_md_meth_names = None # (!) real value is ''
| [
"Shreyasrameshkulkarni@gmail.com"
] | Shreyasrameshkulkarni@gmail.com |
4d19c858d848ed41fa60d015b55f3e9820b42186 | 926ec284a08a72964fb81a68e342d09f33fa6d3f | /omaha_server/omaha/parser.py | f2a84365bee435000ca942451c8a13819cc0f089 | [
"Apache-2.0",
"LicenseRef-scancode-warranty-disclaimer"
] | permissive | yukels/omaha-server | a2395431978e17174fe76513b6cada99cf5317eb | 48406bef1e77e393031e410454a6ee10e8c47925 | refs/heads/master | 2021-01-17T23:53:03.781282 | 2016-07-25T14:00:55 | 2016-07-25T14:00:55 | 55,139,490 | 0 | 0 | null | 2016-03-31T09:53:58 | 2016-03-31T09:53:58 | null | UTF-8 | Python | false | false | 3,031 | py | # coding: utf8
"""
This software is licensed under the Apache 2 license, quoted below.
Copyright 2014 Crystalnix Limited
Licensed under the Apache License, Version 2.0 (the "License"); you may not
use this file except in compliance with the License. You may obtain a copy of
the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
License for the specific language governing permissions and limitations under
the License.
"""
import os
from lxml import etree, objectify
__all__ = ['parser', 'parse_request']
BASE_DIR = os.path.dirname(__file__)
with open(os.path.join(BASE_DIR, 'request.xsd')) as f:
schema = etree.XMLSchema(file=f)
parser = objectify.makeparser(schema=schema)
def parse_request(request):
"""
>>> request = b'''<?xml version="1.0" encoding="UTF-8"?>
... <request protocol="3.0"
... version="1.3.23.0"
... ismachine="0"
... sessionid="{5FAD27D4-6BFA-4daa-A1B3-5A1F821FEE0F}"
... userid="{D0BBD725-742D-44ae-8D46-0231E881D58E}"
... installsource="scheduler"
... testsource="ossdev"
... requestid="{C8F6EDF3-B623-4ee6-B2DA-1D08A0B4C665}">
... <os platform="win" version="6.1" sp="" arch="x64"/>
... <app appid="{430FD4D0-B729-4F61-AA34-91526481799D}" version="1.2.23.0" nextversion="" lang="en" brand="GGLS"
... client="someclientid" installage="39">
... <updatecheck/>
... <ping r="1"/>
... </app>
... <app appid="{D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C}" version="2.2.2.0" nextversion="" lang="en" brand="GGLS"
... client="" installage="6">
... <updatecheck/>
... <ping r="1"/>
... </app>
... </request>'''
>>> request_obj = parse_request(request)
>>> request_obj.get('version')
'1.3.23.0'
>>> request_obj.os.get('platform')
'win'
>>> request_obj.app.get('appid')
'{430FD4D0-B729-4F61-AA34-91526481799D}'
>>> request_obj.app.find('updatecheck')
''
>>> request_obj.keys()
['protocol', 'version', 'ismachine', 'sessionid', 'userid', 'installsource', 'testsource', 'requestid']
>>> request_obj.values()
['3.0', '1.3.23.0', '0', '{5FAD27D4-6BFA-4daa-A1B3-5A1F821FEE0F}', '{D0BBD725-742D-44ae-8D46-0231E881D58E}', 'scheduler', 'ossdev', '{C8F6EDF3-B623-4ee6-B2DA-1D08A0B4C665}']
>>> request_obj.tag
'request'
>>> for app in request_obj.find('app'):
... app.get('appid')
...
'{430FD4D0-B729-4F61-AA34-91526481799D}'
'{D0AB2EBC-931B-4013-9FEB-C9C4C2225C8C}'
"""
return objectify.fromstring(request, parser)
| [
"yurtaev.egor@gmail.com"
] | yurtaev.egor@gmail.com |
6479a595ec5e5e6a86e7178104d6df7763bfa983 | 5f58a50d7c44d0cf612b9076df40da89302b5ba6 | /geeadd/batch_copy.py | ff05da1fea9836b073de8a843dc1faa2c53b45c2 | [
"Apache-2.0"
] | permissive | jkrizan/gee_asset_manager_addon | 386a2a5b96e31bdb5e40a08ad12545e11a376764 | 884793185ef5641f0b53349feb5f4c3be272fd28 | refs/heads/master | 2020-05-19T12:58:15.830923 | 2019-01-01T16:46:16 | 2019-01-01T16:46:16 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 716 | py | from __future__ import print_function
import ee
import os
ee.Initialize()
def copy(collection_path,final_path):
assets_list = ee.data.getList(params={'id': collection_path})
assets_names = [os.path.basename(asset['id']) for asset in assets_list]
print('Copying a total of '+str(len(assets_names))+'.....')
for count,items in enumerate(assets_names):
print ('Copying '+str(count+1)+' of '+str(len(assets_names)), end='\r')
init=collection_path+'/'+items
final=final_path+'/'+items
try:
ee.data.copyAsset(init,final)
except Exception as e:
pass
#batchcopy(collection_path='users/samapriya/Belem/BelemRE',final_path='users/samapriya/bl')
| [
"samapriya.roy@gmail.com"
] | samapriya.roy@gmail.com |
89c371ab6dc5ad908b6e19a063cb752e11d707e5 | fe8f10e3333954a09d135e159efbf161365033df | /myAssembler | d9cf64bbc2b2d6cc17ce8d5d18b1e2f1d4c29440 | [] | no_license | akachigerry/gerry-repo | b02973d3030cfbbe1f7144c6a6ddfa3b1b136095 | 91b9963c61b7aa97a4aa21f5186390e273e6cfe9 | refs/heads/master | 2021-01-15T14:50:09.511895 | 2018-08-12T05:34:52 | 2018-08-12T05:34:52 | 99,695,696 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 22,063 | #!/usr/bin/python
import sys, getopt
import itertools
from itertools import imap, starmap, izip
from multiprocessing import Pool
from multiprocessing import Process, Queue, cpu_count, sharedctypes, Manager
from multiprocessing.dummy import Pool as ThreadPool
import inspect
import random
import counter
import re
import operator
from ctypes import *
from multiprocessing.sharedctypes import Array
from collections import deque
def input_arg(argv):
FastqFile = ''
WordSize = ''
try:
opts, args = getopt.getopt(argv,"hf1:f2:k:rl:")
except getopt.GetoptError, err:
print str(err)
usage()
sys.exit(2)
for opt, arg in opts:
if opt == '-h':
print 'my_Assembler.py -f1 <FastqFile1> -f2 <FastqFile2> -k <WordSize>'
sys.exit()
elif opt in ("-f1"):
FastqFile1 = arg
elif opt in ("-f2"):
FastqFile2 = arg
elif opt in ("-k"):
WordSize = arg
input_arg(sys.argv[1:])
linecnt = 0
qualst = []
fastq_inp = open(sys.argv[2])
readcnt = 0
#define a dictionary with reads as keys and count of read occurences as values
reads=dict()
k=int(sys.argv[6])
#define a node class which holds the sink(s) of a source.
#there are two label attributes for this class because no source should
#have more than two sinks. This requirement is necessary in order to
#simplify the task of finding the Eulerian path.
class Node:
def __init__(self, par):
self.label1 = par
self.label2 = ''
self.indegree1 = 0
self.indegree2 = 0
def prep_reads(reads):
readlst = []
for i in xrange(len(reads)):
read = reads[i].rstrip('\n')
if ''.join(sorted(set(list(read)))) == 'ACGT':
readlst.append(read)
continue
if ''.join(sorted(set(list(read)))) == 'ACGNT':
read1 = ''
for i in xrange(len(read)):
if read[i] == 'N':
read1+='A'
else:
read1+=read[i]
readlst.append(read1)
return readlst;
def read_index_pair(read):
looplen = len(read)-(k+1)
loop_range = range(looplen)
readlst = [read]*looplen
return zip(readlst, loop_range);
def create_words(read):
looplen = len(read)-(k+1)
return [(read[x:x+k], read[x+1:x+1+k]) for x in xrange(looplen)];
#round count of outgoing edges per source word to the nearest multiple of 5.
def round_mult5(value):
cnt = int(5 * round(float(value)/5))
return cnt;
#generate freq of use numbers of source-to-sink outword connections, counted above.
#This is necessary to estimate the least connection size which is most frequent.
def edge_freq(x):
return x.indegree;
def class2list(x):
if isinstance(entire_graph[x], list) == False:
entire_graph[x] = [entire_graph[x]]
def suffixArray(s):
#Given T return suffix array SA(T). We use Python's sorted
#function here for simplicity, but we can do better.
# Empty suffix '' plays role of $.
satups = sorted([(s[i:], i) for i in xrange(0, len(s)+1)])
# Extract and return just the offsets
return map(lambda x: x[1], satups)
def bwt(t):
#Given T, returns BWT(T), by way of the suffix array.
bw = []
for si in iter(suffixArray(t)):
if si == 0:
bw.append('$')
else:
bw.append(t[si-1])
return bw;
def rankBwt(bw):
#Given BWT string bw, returns a parallel list of B-ranks. Also
#returns tots, a mapping from characters to # times the
#character appears in BWT.
tots = dict()
seen = dict()
grp = 0
lastchar = ''
for c in iter(bw):
if c not in seen.keys():
grp=1
seen[c] = grp
tots[c+str(grp)] = 1
lastchar = c
if c in seen.keys() and c == lastchar:
tots[c+str(grp)]+=1
lastchar = c
if c in seen.keys() and c != lastchar:
grp = seen[c]+1
tots[c+str(grp)] = 1
lastchar = c
seen[c]+=1
return tuple(tots.items());
def firstCol(tots):
#Return a map from characters to the range of cells in the first
#column containing the character.
first = {}
startpos = 0
for c, count in sorted(iter(tots)):
if c[0] in first.keys():
startpos+=count
if c[0] not in first.keys():
first[c[0]] = startpos
startpos+=count
return first
def reverseBwt(bw, first):
rowi = 0
t = "$"
seen = ''
bwranks = []
for c in iter(bw):
if c in seen:
rank = seen.count(c)
bwranks.append((c, rank))
seen+c
if c not in seen:
rank = 0
bwranks.append((c, rank))
seen+c
while bw[rowi][1] != '$':
c = bw[rowi]
t = c + t
rowi = first[c]+bwranks[rowi][0]
return t;
def uncompress(tup):
#step1:reproduce a bwt string from the compressed dictionary form
src_bw = ''
for innertup in iter(tup):
c = innertup[0][0]*innertup[1]
src_bw += c
#step2: reproduce original sting from bwt string
first_dic = firstCol(tup)
string = reverseBwt(src_bw, fisrt_dic)
return string;
#next, generate contig
def contig_gen(source):
#start building contigs from source strings with only outgoing edges
contig = source
current = source
for x in xrange(len(graph.keys())):
if current in graph.keys():
#pick next node(s) to be traversed (label1, then label2)
if len(graph[current].label1) > 0:
nxt = graph[current].label1
else:
print contig
return contig;
#test if current could be start of a repeat loop by checking if it has
#greater than or equal to two edge unit connections to its label1 sink
looptest = 0
if int(round(graph1[current].indegree1/depth)) >= 2:
looptest+=1
break
if looptest == 0:
#check label1 (i.e nxt) of the source 'current' for potential of being a source.
#This potential depends on the sink having outgoing connection to sink(s).
newcurrent = 0
if nxt in graph.keys():
contig+=nxt[-1]
graph[current].label1 = ''
current = nxt
nxt = graph[current].label1
continue
else:
print contig
return contig;
if looptest > 0:
#check if the source word named 'current'
#is the start of a repeat loop block. Do this by checking if number
#of its connection to its first sink is greater than double the depth.
loopstart = current
cnt = 0
repeatcnt = 0
#create a loop through source-sink edges in order to confirm the repeat sequence.
#The number of passes through the loop should be determined from the
#max number of repeat monomers(N) obtainable in a plant or animal genome
#and the k-word length (k) as follows: N-(k+2). The result is approximated to the
#closest higher integer.
yesRepeat = 0
for x in xrange(400):
if nxt in graph.keys():
#if loopstart is equal to nxt of next sink
if loopstart == graph[nxt].label1:
yesRepeat+=1
contig+=nxtlst[i][-1]
source = nxt
graph[current].label1=''
numRepeat = int(round(graph[source].indegree1/depth))+1
contig = contig*numRepeat
#delete the 'loopstart' sink of nxt before continuing contig extension from main loop.
#this will prevent potential of unnecessarily trasversing the repeat monomer again given that
#the entire repeat sequence has already been estimated from its monomer.
current = source
nxt = graph[current].label1
break
else:
print contig
return contig;
if yesRepeat == 0:
#else, make the first sink which is in turn a source of some other sink(s),
#the new current and continue through the (400X) loop.
if nxt in graph.keys():
if len(graph[nxt].label1) > 0:
contig+=nxt[-1]
graph[current].label1=''
current = nxt
else:
print contig
return contig;
else:
print contig
return contig;
return contig;
def tuple2dict(item):
dic={}
source = item[0][0]
sink = item[0][1]
dic.setdefault(source, {})[sink]=item[1]
return dic;
def wordCompress(tup):
source_bwt = bwt(tup[0])
source_bwt = rankBwt(source_bwt)
sink_bwt = bwt(tup[1])
sink_bwt = rankBwt(sink_bwt)
compdict[(source_bwt, sink_bwt)] = wordlst[tup]
#define function to get input from queue and put output to it.
def fun(f, X, q_in, q_out):
while True:
itm = q_in.get()
if itm is None:
break
q_out.put(f(itm))
def parmap(f, X, nprocs=cpu_count()):
q_in = Queue()
q_out = Queue()
proc = [Process(target=fun, args=(f, X, q_in, q_out)) for n in xrange(nprocs)]
for n in xrange(nprocs):
proc[n].daemon = True
proc[n].start()
sent = list(q_in.put(itm) for itm in X)
[q_in.put(None) for i in xrange(nprocs)]
res = list(q_out.get() for i in xrange(len(sent)))
[proc[n].join() for n in xrange(nprocs)]
print len(res)
return res;
#instance of class takes in a generator (which yields k, v tuples) and builds a dictionary-like class object on the fly.
#This is a more memory-efficient way of building a dictionary
class LazyDict():
"""A dictionary built on demand from an iterator."""
def __init__(self, iterator):
self._dict = dict(iterator)
self._iterator = iterator
def __getitem__(self, key):
if key in self:
return self._dict[key]
else:
raise KeyError(key)
def __setitem__(self, key, value):
return self._dict__setitem__(key, value)
def keys(self):
return self._dict.keys()
def iterkeys(self):
return self._dict.iterkeys()
def values(self):
return self._dict.values()
def itervalues(self):
return self._dict.itervalues()
def items(self):
return self._dict.items()
def iteritems(self):
return self._dict.iteritems()
def update(self, iterator):
return self._dict.update(self._iterator)
#__contain__ is an abstract base class implementation of
#'in' used in __getitem__ method above
def __contains__(self, key):
if key in self._dict:
return True
else:
return False
fastq_inp1 = open(sys.argv[2])
fastq_inp2 = open(sys.argv[4])
fastq1 = itertools.islice(fastq_inp1, 1, None, 4)
fastq2 = itertools.islice(fastq_inp2, 1, None, 4)
del fastq_inp1, fastq_inp2
fastqall = []
fastqall.extend(fastq1)
fastqall.extend(fastq2)
print len(fastqall)
del fastq2, fastq1
chunksize = int(round(len(fastqall)/cpu_count()))
print chunksize
print cpu_count()
def chunking(lst):
z=0
chunklst = []
for i in xrange(cpu_count()+1):
chunk = list(itertools.islice(lst[z:], chunksize))
chunklst.append(chunk)
z+=chunksize
if z >= len(lst) - chunksize:
if z >= len(lst):
return chunklst;
else:
chunklst.append(lst[z:])
return chunklst;
return chunklst;
fastqall_chunk = chunking(fastqall)
print len(fastqall_chunk[0])
print len(fastqall_chunk)
print sys.getsizeof(fastqall_chunk)
pool = Pool(cpu_count())
wordlst=[]
for result in pool.imap(prep_reads, fastqall_chunk):
wordGen = (create_words(read) for read in result)
resultLen=len(result)
print resultLen
for i in xrange(resultLen):
wordlst.extend(wordGen.next())
pool.close()
pool.join()
del fastqall_chunk
#define the function 'trim' to remove source words having total of 2 or less edge connections to a sink
def trim(itm):
if itm[1] > 2:
return itm;
print "size of wordlst is", sys.getsizeof(wordlst)
#break fastq files into chunks, then loop throuugh list of chunks creating words for each loop.
#a dynamic chunking is utilised to maximize available system memory and processes for speedy execution.
wordlst=counter.Counter(wordlst).most_common()
print wordlst[0:3]
pool = Pool(20)
wordlst1 = pool.map(trim, wordlst)
pool.close()
pool.join()
print len(wordlst1)
del wordlst
src2sink_cnt = []
allsinks = []
graph = {}
looptracker=0
for i in xrange(len(wordlst1)):
looptracker+=1
item = wordlst1[i]
if item != None:
source = item[0][0]
sink = Node(item[0][1])
if source not in graph.keys():
allsinks.append(item[0][1])
sink.indegree1 += int(item[1])
graph[source] = sink
src2sink_cnt.append(int(item[1]))
continue
else:
if graph[source].label2 == '':
allsinks.append(item[0][1])
graph[source].label2 += item[0][1]
graph[source].indegree2 += int(item[1])
src2sink_cnt.append(int(item[1]))
else:
print 'maximum number(2) of unique sinks allowed for each source has been reached'
print graph.items()[0:3]
print len(src2sink_cnt)
def merge_dic(diclst):
dic=dict()
cnt=0
for d in diclst:
if cnt == 0:
dic.update(d)
else:
dickeys = dic.keys()
next_dickeys = d.keys()
common_dickeys = list(set(dickeys).intersection(set(next_dickeys)))
diff_dickeys = list(set(dickeys).difference(set(next_dickeys)))
if len(common_dictkeys) == 0:
dic.update(d)
else:
for key in iter(diff_dickeys):
unitdic = {key:diff_dickeys[key]}
dic.update(unitdic)
#for set of keys common between new dictionary and current merged dictionary
for key in iter(common_dictkeys):
next_val_of_val = d[key].values()
next_key_of_val = d[key].keys()
if next_value != dic[key]:
graph2[tupl]+=next_value
#generate a dictionary from the list of source-sink tuples(wordlst)
#the list should have source strings as key. Values for each of these
#must hold all unique sinks of a source
#To fulfill the above, define a function holds the tuple in
#a dictionary as follows:
#source as main dictionary key
#all sinks of the source as a sub-dictionary consisting of each unique
#sink of the source as key and a number, reflecting the sink's frequeny
#as value.
chunksize = int(round(len(src2sink_cnt)/cpu_count()))
pool = Pool(cpu_count())
src2sink_cnt = pool.imap_unordered(round_mult5, src2sink_cnt, chunksize)
pool.close()
pool.join()
src2sink_cnt = counter.Counter(src2sink_cnt)
print len(src2sink_cnt)
kmatchFreq_dict = LazyDict(src2sink_cnt)
#sort above dictionary
kmatchFreq_dict = sorted(kmatchFreq_dict.items(), key=operator.itemgetter(1), reverse=True)
print kmatchFreq_dict
#The most frequent kword matches (number of source-to-sink connections)
#should represent the approximate sequencing depth.
kmatchFreq_dict = LazyDict(src2sink_cnt)
depth = kmatchFreq_dict[0][0]
del src2sink_cnt
print depth
#Now to the next stage: trace a Eularian path through the graph
#first off, generate list of contig start strings which will be used as input.
#after removing edges potentially arising from seq error (trim function),
#source words with more than one unique sink could result from one of two situations:
#1. existence of alleles of same locus
#2. existence of genome segment copies or paralogs on different loci of same chromosome
#the first case is not tolerable for contig generation as it could result in contigs
#containing homologous haplotype segments in adjacent positions. For this reason,
#the graph is prunned to remove instances of the first case.
#the second case is tolerable iff the difference between indegree and outdegree for the node is not greater than unity (1).
cnt = 0
cntmin = -1
prevkeylst = []
startsources = deque()
secondBranchStartList = []
graphkeys=graph.iterkeys()
for key in graphkeys:
#define a 'futre key i.e the sink to the current source or the next source after the current (futkey)'
#to allow a more accurate check of whether node split points (outgree - indgree =1)
#indicate potential heterozygotes or not
futkey = graph[key].label1
src_in_sinkcnt = allsinks.count(futkey)
src_in_sinkcnt1 = allsinks.count(key)
if futkey in graph.keys():
if graph[futkey].label2 != '':
currentEdge = graph[key].indegree1
futureEdge1 = graph[futkey].indegree1
futureEdge2 = graph[futkey].indegree2
#to check if a source connection to multiple sinks is due to het alleles:
#1. their must be 2 sinks connected to the source
#2. number of edges (connections) to the source must be approximately equal to sum for both sinks
#3. number of edges must be approximately half of the estimated depth
if int(round(futureEdge1/graph[futkey].indegree2)) == 1 and \
int(currentEdge/futureEdge1) >= 1.5 and \
int(currentEdge/futureEdge1) <= 2.5:
#the seconnd sink is eliminated if above test for heterozygosity is confirmed. While not useful for strict genome assembly as here,
#the above heterozygosity test would be invaluable for assemmbly of haplotypes and for reference-free variant extraction.
graph[futkey].label2 = ''
#Define potential contig starts. These would be source words not found in sinklist and
#label2 of branch nodes. This label2 must however be a source itself
if (src_in_sinkcnt1 == 0 and graph[key].label2 == '') or (src_in_sinkcnt1 == 1 and graph[key].label2 != ''):
#use appendleft property of deque to ensure that the starting word for the genome (source not found in sink list)
if (src_in_sinkcnt1 == 0 and graph[key].label2 == ''):
startsources.appendleft(key)
if (src_in_sinkcnt1 == 1 and graph[key].label2 != ''):
if graph[key].label2 in graph.keys():
startsources.append(graph[key].label2)
secondBranchStartList.append(graph[key].label2)
del graphkeys
print len(startsources)
#define function to put input data into queue, define specific number
#of processes for analysing each component of this data
#(according to a pre-defined function) and output the result in queue
contiglst = []
pool=Pool(cpu_count())
if cpu_count() >= len(startsources) or int(round(len(startsources)/cpu_count())) < 2:
for result in pool.imap_unordered(contig_gen, startsources):
contigs =[result, len(result)]
contiglst.append(result)
pool.close()
pool.join()
if int(round(len(startsources)/cpu_count())) >= 2:
for result in pool.imap_unordered(contig_gen, startsources, int(round(len(startsources)/cpu_count()))):
contigs = [result, len(result)]
contiglst.append(result)
pool.close()
pool.join()
print "length of contiglst before contig lengthning", len(contiglst)
#go through the contig list and search for branching nodes(sources) and use these as
#inter-contig connections to lengthen contigs
newcontiglst=[]
allcontig = '_'.join(contiglst)
for i in xrange(len(secondBranchStartList)):
longcontig = ''
#branchnode will be the entire length of a secondBranchStartList item except for last base.
#this strategy will ensure that the source word for the branching label2 sinks, from which secondBranchStartList items
#were populated, can be found and the branch contig extended toward the 5' direction.
branchnode = secondBranchStartList[i][0:-1]
if branchnode in allcontig:
branchstart = allcontig.find(branchnode)
#avoid matches at start of contig as they are branch node. interest is
#in longer contigs that incorporate source of branch node start (label2)
#somewhere along its length.
if branchstart == 0:
newcontiglst.append(secondBranchStartList[i])
continue
addcontig = allcontig[0:branchstart-1].split('_')[-1]
longcontig = addcontig + secondBranchStartList[i]
newcontiglst.append(longcontig)
else:
newcontiglst.append(secondBranchStartList[i])
with open("contig_out2", 'w') as outfile:
contiglen=[]
for itm in newcontiglst:
outfile.write('\t'.join([itm, len(itm)]))
contiglen.append(len(itm))
#calculate N50
halfassem = int(round(sum(contiglen)/2))
n50 = 0
contiglist = sorted(contiglen, reverse=True)
for num in iter(contiglist):
if n50 < halfassem:
n50+=num
else:
break
print "N50 =", n50
| [
"noreply@github.com"
] | noreply@github.com | |
f254c37980b3283334624c8365ad413025cf7c98 | 3cafdf8967c5e1c331a312684e550353cf10fffd | /Older_versions/NCBI.py | 5d0dea2e726821c3cf8f724948ef928a8cd3cbc4 | [] | no_license | TiagoS16/SARS-CoV-2-e-genes | 27a516a48dd84f4b0b5cf28a7f5ed417c92da2a2 | f301595abcc1258639d9ecf85df778d6f9cb7bee | refs/heads/master | 2023-07-19T21:41:53.431125 | 2021-08-25T15:59:29 | 2021-08-25T15:59:29 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 545 | py | from Bio import Entrez
from Bio import SeqIO
word = 'FGB[Gene]'
res= 1
email= 'pg42877@alunos.uminho.pt'
Entrez.email= email
handle = Entrez.esearch(db = 'nucleotide', term=word, retmax= res)
record = Entrez.read(handle)
gi_list = record['IdList']
print(gi_list)
for a in gi_list:
a = 'NG_008833.1'
handle = Entrez.efetch(db="nucleotide", id=a, rettype="gb", retmode="text")
record = SeqIO.read(handle, "genbank")
save_file = open('my_blast.xml', 'w')
save_file.write(handle.read())
handle.close()
print(record.id)
| [
"74024284+MiguelBarros99@users.noreply.github.com"
] | 74024284+MiguelBarros99@users.noreply.github.com |
b6979f9099ddfa64a87e7398c649f7d22aafcedc | 603954bafbe9331e5f07b447b1147887ed8c9c39 | /polls/serializers.py | bb29f2e597cdcad26f56d70e2a169702e5c71bac | [] | no_license | tanmayawasekar/python-practice | 8bc8990e03805a532bed9939edbf4d65cbd106de | 19ed39a23f4fff64a53c66a0580791e0bebf894e | refs/heads/master | 2021-04-26T23:57:07.090753 | 2018-03-29T06:46:56 | 2018-03-29T06:46:56 | 123,884,568 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 227 | py | from .models import Choice
from rest_framework import serializers
class UserSerializer(serializers.HyperlinkedModelSerializer):
class Meta:
model = Choice
fields = ('choice_text', 'integer_field', 'votes')
| [
"tanmay.awasekar@coverfoxmail.com"
] | tanmay.awasekar@coverfoxmail.com |
82d31d3429aec8f65ca302385483553480dfd265 | a57fa788fc8d4ace67c42c4d4ea7083bf139493a | /test/win_test_disk_drive.py | d4d6fa43f1088c427bc714d9087f573a0a740c29 | [
"MIT",
"LicenseRef-scancode-unknown-license-reference"
] | permissive | venanciofire/py_win_unc | 9a221b175f0f4e0ba66e2f9f632bee8aaf5ff6b6 | e46256560cb6acfdbfbe8992d454e7ea492cb6ba | refs/heads/master | 2023-03-22T07:09:04.640066 | 2013-09-30T15:58:23 | 2013-09-30T15:58:23 | 588,721,538 | 1 | 0 | MIT | 2023-01-13T20:34:25 | 2023-01-13T20:34:25 | null | UTF-8 | Python | false | false | 354 | py | from unittest import TestCase
from win_unc.errors import NoDrivesAvailableError
from win_unc.disk_drive import get_available_disk_drive
class TestAvailableDiskDrive(TestCase):
def test_get_available_disk_drive(self):
try:
self.assertIsNotNone(get_available_disk_drive())
except NoDrivesAvailableError:
pass
| [
"elliot@3noch.com"
] | elliot@3noch.com |
64e20e5645269edc49ca0d18451cd287e6d86f65 | 7ff9de453f53b658d13896bb0c376d67489145a7 | /logics/exercise.py | 2a82386f9fc82b4d34d67b3a2bfb7ac928f3b22d | [
"MIT"
] | permissive | danielkpodo/python-zero-to-mastery | aa4851fd0bfe1f0bfa4191fa141fa9551fd7c7a9 | d39468f48211bc82e4e2613745d9107d433e05af | refs/heads/master | 2022-11-21T10:55:54.776490 | 2020-07-19T15:07:35 | 2020-07-19T15:07:35 | 275,909,257 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 341 | py | # sum up the total of this list using looping
from functools import reduce
items = [1, 3, 4, 5, 6, 7, 7, 5, 7, 6, 6, 3]
total = 0
for num in items:
total += num
print(total)
print(sum(items))
# Higher order functions
def summer(accumulator, initial):
return accumulator + initial
result = reduce(summer, items)
print(result)
| [
"kpododanielnarh@gmail.com"
] | kpododanielnarh@gmail.com |
9c7e3a54253670e920ac976c5c6efa681e7cb585 | 37f19a6a4369b5527253ffe350b9e8263f151f1e | /cDCheck/cDCheck.py | 6c7e01372209eecbe43ba5f234de7c671abf15e8 | [
"MIT"
] | permissive | csm10495/cDCheck | a63897a5665605e5da7622e6ec9d96d31059cdc6 | 5e19c938ad78a79dfad6f4a78fdff8c557dc49ce | refs/heads/master | 2021-01-23T06:54:20.623379 | 2015-05-03T20:13:37 | 2015-05-03T20:13:37 | 34,648,930 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 5,304 | py | #cDCheck: A Python script to check for, and delete duplicate files in a directory
#(C) Charles Machalow - MIT License
import os #for directory access
import sys #for args
import threading #for threading
import re #for regex matching
#processes the files in range
def processRange(r1, r2, file_dict, dup_file_dict, files):
for i in range(r1, r2):
#hash as binary
h = hash(open(files[i], "rb").read())
if h in file_dict:
#print("adding to dup")
if h in dup_file_dict:
dup_file_dict[h].append(files[i])
else:
dup_file_dict[h] = [files[i], file_dict[h]]
else:
file_dict[h] = files[i]
#print("adding to file_dict")
#alerts the user to duplicates
def callOutDups(dup_file_dict):
for i in dup_file_dict:
print("Duplicate file detected with hash: " + str(i))
print("Instances:")
for j, k in enumerate(dup_file_dict[i]):
print(str(j) + ": " + str(k))
#keep going to valid input
while True:
c = input("Choose a number for the file you would like to maintain. Other options are:\ns to skip this file\nr to delete all files that DON'T contain a regex match\n")
#break character
if str(c).lower() == "s":
break
#regex character
if str(c).lower() == "r":
r = input("Regex: ")
try:
reg = re.compile(str(r))
except Exception:
print("Unable to compile regex. Please try again.")
continue
for z in range(0, j + 1):
#delete all that don't match regex
if not reg.search(dup_file_dict[i][z]):
os.remove(dup_file_dict[i][z])
print("Deleted files that didn't match regex: " + str(r))
break
try:
c = int(c)
except ValueError:
print("Invalid input, choose one file (by number) to maintain")
continue
#make sure given int is valid
if c >= 0 and c <= j:
print("Performing requested action. Maintaining file " + str(c) + ". Deleting others.")
for z in range(0, j + 1):
if z != c:
os.remove(dup_file_dict[i][z])
break
else:
print("Invalid input, choose one file (by number) to maintain")
#does the iteration work
def checkPath(path, thread_count=4):
file_dict = {}
dup_file_dict = {}
file_count = 0
files = []
print("Processing files in directory: " + path)
ldir = os.listdir(path)
for i in ldir:
f_path = os.path.join(path, i)
if os.path.isfile(f_path):
file_count+=1
files.append(f_path)
print("Files found: " + str(file_count))
threads = []
f_slice = []
#handle if we can do more threads than files
if (thread_count > file_count):
thread_count = file_count
#starting per thread
per_thread = int(file_count / thread_count)
#set all threads
for i in range(thread_count):
f_slice.append(per_thread)
#remainder number of files that haven't been distributed to threads
extra_files = file_count - (per_thread * thread_count)
#add remainder to threads as equally as possible
for i in range(len(f_slice)):
if extra_files == 0:
break
f_slice[i]+=1
extra_files -= 1
#starts a thread_count threads
#fill threads list with threads that we can start
#f_slice is the number of files each thread should hash
counter = 0
for i in range(len(f_slice)):
s1 = counter
counter = counter + f_slice[i]
t = threading.Thread(target=processRange, args=(s1, counter, file_dict, dup_file_dict, files))
threads.append(t)
#start all threads
for i in threads:
i.start()
#join all threads
for i in threads:
i.join()
print("Done Processing Directory\n")
print("Found " + str(len(dup_file_dict)) + " files with duplicates")
callOutDups(dup_file_dict)
#entrant function
def main():
#make sure we have enough args
if len(sys.argv) >= 2:
print("Please do not remove files from the given directory while this is running")
path = sys.argv[1]
#make sure path exists
if os.path.exists(path):
if (len(sys.argv) == 3):
try:
t = int(sys.argv[2])
checkPath(path, t)
print("cDCheck Completed Successfully!")
except ValueError:
print("Number of threads is not an integer, please make it one and try again")
if (len(sys.argv) == 2):
checkPath(path)
print("cDCheck Completed Successfully!")
else:
print("Given path does not exist, please check and try again")
else:
print("Usage: python cDCheck.py folderpath <number of threads, defaults to 4>")
if __name__ == '__main__':
main() | [
"csm10495@gmail.com"
] | csm10495@gmail.com |
ab69c6dd82887d2be098cbc3924977fc48cae5df | b367245af0a5e4284afab9f7177c5b136b41283c | /swa-paddle/models/vgg.py | a6858c389dafae27dc7b8100769bfad2f5e111f5 | [
"BSD-2-Clause"
] | permissive | ncpaddle/SWA-paddle | 0544c38a1eb50c9bcf0db460b91a4f0c7beec48c | 48c2642294062e2cf6f0d21c5a1cd30e1002b1f1 | refs/heads/master | 2023-08-24T02:09:06.385258 | 2021-11-03T02:05:35 | 2021-11-03T02:05:35 | 419,941,948 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,386 | py | """
VGG model definition
ported from https://github.com/pytorch/vision/blob/master/torchvision/models/vgg.py
"""
import math
import paddle
import paddle.nn as nn
from paddle.vision import transforms
__all__ = ['VGG16', 'VGG16BN', 'VGG19', 'VGG19BN']
def make_layers(cfg, batch_norm=False):
layers = list()
in_channels = 3
for v in cfg:
if v == 'M':
layers += [nn.MaxPool2D(kernel_size=2, stride=2)]
else:
conv2d = nn.Conv2D(in_channels, v, kernel_size=3, padding=1, weight_attr=nn.initializer.KaimingNormal())
if batch_norm:
layers += [conv2d, nn.BatchNorm2D(v), nn.ReLU()]
else:
layers += [conv2d, nn.ReLU()]
in_channels = v
return nn.Sequential(*layers)
cfg = {
16: [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 'M', 512, 512, 512, 'M', 512, 512, 512, 'M'],
19: [64, 64, 'M', 128, 128, 'M', 256, 256, 256, 256, 'M', 512, 512, 512, 512, 'M',
512, 512, 512, 512, 'M'],
}
class VGG(nn.Layer):
def __init__(self, num_classes=10, depth=16, batch_norm=False):
super(VGG, self).__init__()
self.features = make_layers(cfg[depth], batch_norm)
self.classifier = nn.Sequential(
nn.Dropout(),
nn.Linear(512, 512, bias_attr=True),
nn.ReLU(True),
nn.Dropout(),
nn.Linear(512, 512, bias_attr=True),
nn.ReLU(True),
nn.Linear(512, num_classes, bias_attr=True),
)
def forward(self, x):
x = self.features(x)
x = paddle.reshape(x, [x.shape[0], -1])
# x = x.view(x.size(0), -1)
x = self.classifier(x)
return x
class Base:
base = VGG
args = list()
kwargs = dict()
transform_train = transforms.Compose([
transforms.RandomHorizontalFlip(),
transforms.RandomCrop(32, padding=4),
transforms.ToTensor(),
transforms.Normalize((0.485, 0.456, 0.406), (0.229, 0.224, 0.225))
])
transform_test = transforms.Compose([
transforms.ToTensor(),
transforms.Normalize((0.485, 0.456, 0.406), (0.229, 0.224, 0.225))
])
class VGG16(Base):
pass
class VGG16BN(Base):
kwargs = {'batch_norm': True}
class VGG19(Base):
kwargs = {'depth': 19}
class VGG19BN(Base):
kwargs = {'depth': 19, 'batch_norm': True}
| [
"845749146@qq.com"
] | 845749146@qq.com |
9cd9d566ee77d3702c5658462da130d13dcc8e87 | 2789c64d06b08722528c4ee22169020dc7c22c4d | /karalamalar_src_idi/ilkmodul.py | d56052a0ccf400f71faa2c896129d55074e74a1d | [] | no_license | kadirkoyici/pythonkursu | d209f992a50c650cc6c79132a9ec41f0e4bd8f81 | f425a908353749791480624cbb26e1a409e805bb | refs/heads/master | 2021-01-19T10:10:58.906420 | 2015-03-09T06:50:32 | 2015-03-09T06:50:32 | 28,480,514 | 0 | 1 | null | 2015-01-06T08:44:51 | 2014-12-25T11:57:39 | Python | UTF-8 | Python | false | false | 179 | py | #!/usr/bin/env python
#-*- coding:utf-8 -*-
program = "Python"
surum = "2.x"
def faktoriyel(sayi):
fak = 1
for i in range(sayi):
fak= fak * (i+1)
return fak | [
"kadirkoyici@hotmail.com"
] | kadirkoyici@hotmail.com |
e55b057232bdc2f6da6ef749bc1f2f138a2fdb19 | d3e7b50efb9ddfa329eab0411d0f4cf058d4aabe | /tasks/task_c.py | 22a377b7e2147a94a834d1b3c7c12399925dbe2f | [] | no_license | joelbispo/luigi-demo-blackrock | 1a6f774930d71dd9554e453dfc8cc9ab03f656f8 | fdb55df581e79044aced601a96de6edfe0c86382 | refs/heads/master | 2022-11-16T06:49:31.967588 | 2020-07-16T08:38:26 | 2020-07-16T08:38:26 | 280,096,987 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 322 | py | import luigi
from util.cronometro import cronometro
class TarefaC(luigi.Task):
def run(self):
cronometro(5)
with self.output().open('w') as log:
log.write('Oi, Arena, eu terminei')
def output(self):
return luigi.LocalTarget(f'C:\\temp-luigi\\{self.__class__.__name__}.txt') | [
"joelbisponeto@gmail.com"
] | joelbisponeto@gmail.com |
83e185e53ee41e521bdd311be71ebf8b7318349e | 05b8143f004c6531a1d24a66888e2b02a41616cf | /mainApp/apis/cinemas_api.py | 905d41498de23e6efa289decd85035190b6c01d9 | [] | no_license | cangmingssir/flask_tpp | 1b0d8f40fd3298789beffca877874dd45d734987 | e6903a47aa2658a105f79c37a30ef5f44a4d1fab | refs/heads/master | 2020-03-19T12:04:37.056215 | 2018-06-17T08:07:48 | 2018-06-17T08:07:48 | 136,493,916 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,864 | py | # coding:utf-8
from flask import request, session
from flask_restful import Resource, reqparse, fields, marshal_with
from mainApp import dao
from mainApp.models import Cinemas, User, Qx
from mainApp.settings import QX
def check_login(qx):
def check(fn):
def wrapper(*args,**kwargs):
token = request.args.get('token')
if not token:
token = request.form.get('token')
user_id = session.get(token)
loginUser = dao.getById(User,user_id)
if not loginUser:
return {'msg':'请先登录!'}
if loginUser.rights & qx == qx:
return fn(*args,**kwargs)
qxObj = dao.queryOne(Qx).filter(Qx.right==qx).first()
return {'msg':'您没有 {} 权限'.format(qxObj.name)}
return wrapper
return check
class CinemasApi(Resource):
#定义输入字段
parser = reqparse.RequestParser()
parser.add_argument('token')
parser.add_argument('opt',required=True)
parser.add_argument('name',help='电影院名称')
parser.add_argument('city',help='影院城市不能为空')
parser.add_argument('district',help='城市区域不能为空')
parser.add_argument('sort',type=int,default=1)
parser.add_argument('orderby',default='hallnum')
parser.add_argument('limit',type=int,default=10)
parser.add_argument('page',type=int,default=1)
#定义输出字段
cinemas_fields = {
'id':fields.Integer,
'name':fields.String,
'city':fields.String,
'district':fields.String,
'address':fields.String,
'phone':fields.String,
'score':fields.Float,
'hallnum':fields.Integer,
'servicecharge':fields.Float,
'astrict':fields.Integer,
'flag':fields.Boolean,
'isdelete':fields.Boolean
}
out_fields={
'returnValue':fields.Nested(cinemas_fields)
}
def selectCinemas(self,cinemas):
args=self.parser.parse_args()
sort = args.get('sort')
cinemas = cinemas.order_by(('-' if sort ==1 else '')+args.get('orderby'))
pager = cinemas.paginate(args.get('page'),args.get('limit'))
return {'returnValue':pager.items}
@marshal_with(out_fields)
def get(self):
#验证请求参数
args=self.parser.parse_args()
opt =args.get('opt')
city = args.get('city')
district = args.get('district')
#用于查询某城市区域的影城信息
if opt == 'cityAndDistrict':
if city and district:
cinemas=dao.queryOne(Cinemas).filter(Cinemas.city==city,
Cinemas.district==district)
if not cinemas.count():
return {'msg':'该地区没有电影院'}
self.selectCinemas(cinemas)
return {'msg':'城市和城区区域不能为空'}
#用于查询某一城市的影城信息
elif opt == 'city':
if city:
cinemas=dao.queryOne(Cinemas).filter(Cinemas.city==city)
if not cinemas.count():
return {'msg':'该城市没有电影院'}
self.selectCinemas(cinemas)
return {'msg':'搜索城市不能为空'}
#查询所有的影城信息
else:
cinemas=dao.queryAll(Cinemas)
self.selectCinemas(cinemas)
@check_login(QX.DELETE_QX)
def delete(self):
cid = request.args.get('cid')
cinemas = dao.getById(Cinemas,cid)
if not cinemas:
return {'msg':'您删除的影院不存在'}
if not dao.delete(cinemas):
return {'msg':'删除失败'}
return {'msg':'删除成功'}
def post(self):
pass
| [
"mu_tongwu@163.com"
] | mu_tongwu@163.com |
0f49a3e164a082564a9962af1e266ec0ba9b7b33 | 83b7dddc0ff247aae41e332c6c41a5b3dbbcf529 | /src/brewv/bottle.py | 28c0a9ee2d66d00691a21779dd5176affda6b47b | [
"MIT"
] | permissive | lime-green/brew-versions | a453e2891a5b787a0f331e2264020497f78f94dc | 502f4efe56b516df9646533ead745e44bc509ac5 | refs/heads/main | 2023-04-05T14:19:59.854486 | 2021-04-20T16:00:35 | 2021-04-20T16:00:35 | 350,513,815 | 1 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,312 | py | import os
import platform
from requests.exceptions import HTTPError
from .constants import BOTTLE_FILE_SUFFIX, IS_MAC_OS, MAC_VER_TO_CODENAME, SYSTEM
from .util import (
check_output,
download_file,
is_supported_mac_ver,
logger,
mac_ver,
make_request,
)
class BottleNotFound(Exception):
def __init__(self, source_revision=None):
self.source_revision = source_revision
class BottleClient:
def __init__(self, formula_name):
self.formula_name = formula_name
self.bottle_os_identifier = build_bottle_os_identifier()
class GithubBottleClient(BottleClient):
GITHUB_AUTH = {"Authorization": "Bearer QQ=="}
base_url = dict(
Darwin="https://ghcr.io/v2/homebrew/core",
Linux="https://ghcr.io/v2/linuxbrew/core",
)[SYSTEM]
def _parse_manifest(self, manifest_response, ref_name):
source_revision = manifest_response["annotations"].get(
"org.opencontainers.image.revision"
)
for manifest in manifest_response["manifests"]:
annotations = manifest["annotations"]
if annotations["org.opencontainers.image.ref.name"] == ref_name:
return annotations["sh.brew.bottle.digest"], source_revision
return None, source_revision
def download_bottle(self, version, destination_path):
manifest_url = f"{self.base_url}/{self.formula_name}/manifests/{version}"
ref_name = f"{version}.{self.bottle_os_identifier}"
source_revision = None
try:
manifest = make_request(
manifest_url,
headers={
"Accept": "application/vnd.oci.image.index.v1+json",
**self.GITHUB_AUTH,
},
)
bottle_digest, source_revision = self._parse_manifest(manifest, ref_name)
if not bottle_digest:
logger.info("No digest found in manifest")
raise BottleNotFound(source_revision)
bottle_url = (
f"{self.base_url}/{self.formula_name}/blobs/sha256:{bottle_digest}"
)
download_file(
bottle_url,
destination_path,
headers=self.GITHUB_AUTH,
sha256_verification=bottle_digest,
)
except HTTPError:
logger.warning("Got HTTP error when attempting to download bottle")
raise BottleNotFound(source_revision)
class BintrayBottleClient(BottleClient):
base_url = dict(
Darwin="https://homebrew.bintray.com/bottles",
Linux="https://linuxbrew.bintray.com/bottles",
)[SYSTEM]
subject = dict(Darwin="homebrew", Linux="linuxbrew")[SYSTEM]
def download_bottle(self, version, destination_path):
bottle_file_name = (
f"{self.formula_name}-{version}"
f".{self.bottle_os_identifier}.{BOTTLE_FILE_SUFFIX}"
)
url = f"{self.base_url}/{bottle_file_name}"
try:
download_file(url, destination_path)
except HTTPError:
logger.warning("Got HTTP error when attempting to download bottle")
raise BottleNotFound
def download_bottle(formula_name, version, bottle_cache_file):
clients = [GithubBottleClient(formula_name), BintrayBottleClient(formula_name)]
source_revision = None
for client in clients:
logger.info(f"Trying bottle download with {client.__class__.__name__}")
try:
return client.download_bottle(version, bottle_cache_file)
except BottleNotFound as e:
source_revision = e.source_revision
raise BottleNotFound(source_revision)
def build_bottle_os_identifier():
if IS_MAC_OS:
assert is_supported_mac_ver()
codename = MAC_VER_TO_CODENAME[mac_ver()].replace(" ", "_").lower()
if platform.machine() == "arm64":
return f"arm64_{codename}"
return codename
return f"{platform.machine()}_linux"
def get_bottle_cache_location(formula_name, version):
brew_cache_dir = check_output(["brew", "--cache"])
bottle_cache_name = f"{formula_name}--{version}"
return os.path.join(
brew_cache_dir,
f"{bottle_cache_name}.{build_bottle_os_identifier()}.{BOTTLE_FILE_SUFFIX}",
)
| [
"josh.doncastermarsiglio@tophatmonocle.com"
] | josh.doncastermarsiglio@tophatmonocle.com |
eef17bf33c968687f21b8f25b314038cbe056676 | c9923e7aef1ee6df8684a0604114033d0cef6d78 | /cnn3d/training/cutmix.py | 0eb91e5a83c6b1075e0f0c42e984706c39149d1e | [
"MIT"
] | permissive | poissonyzr/DFDC | 4ca1fb936829ddfa411e8d92e36dac3c27f3d75b | b08c31781406e303394642c2af1ab59222115683 | refs/heads/master | 2023-05-13T15:37:58.927522 | 2021-06-01T06:55:00 | 2021-06-01T06:55:00 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,532 | py | import numpy as np
import torch
import torch.nn as nn
import torch.nn.functional as F
def rand_bbox_2d(size, lam):
# lam is a vector
B = size[0]
assert B == lam.shape[0]
W = size[2]
H = size[3]
cut_rat = np.sqrt(1. - lam)
cut_w = (W * cut_rat).astype(np.int)
cut_h = (H * cut_rat).astype(np.int)
# uniform
cx = np.random.randint(0, W, B)
cy = np.random.randint(0, H, B)
#
bbx1 = np.clip(cx - cut_w // 2, 0, W)
bby1 = np.clip(cy - cut_h // 2, 0, H)
bbx2 = np.clip(cx + cut_w // 2, 0, W)
bby2 = np.clip(cy + cut_h // 2, 0, H)
return bbx1, bby1, bbx2, bby2
def rand_bbox_3d(size, lam):
# lam is a vector
B = size[0]
assert B == lam.shape[0]
T = size[2]
W = size[3]
H = size[4]
cut_rat = (1. - lam) ** (1/3.)
cut_t = (T * cut_rat).astype(np.int)
cut_w = (W * cut_rat).astype(np.int)
cut_h = (H * cut_rat).astype(np.int)
# uniform
ct = np.random.randint(0, T, B)
cx = np.random.randint(0, W, B)
cy = np.random.randint(0, H, B)
#
bbt1 = np.clip(ct - cut_t // 2, 0, T)
bbx1 = np.clip(cx - cut_w // 2, 0, W)
bby1 = np.clip(cy - cut_h // 2, 0, H)
bbt2 = np.clip(ct + cut_t // 2, 0, T)
bbx2 = np.clip(cx + cut_w // 2, 0, W)
bby2 = np.clip(cy + cut_h // 2, 0, H)
return bbt1, bbx1, bby1, bbt2, bbx2, bby2
def cutmix_apply(batch, alpha):
batch_size = batch.size(0)
lam = np.random.beta(alpha, alpha, batch_size)
lam = np.max((lam, 1.-lam), axis=0)
index = torch.randperm(batch_size)
if batch.ndim == 5:
# 3D
t1, x1, y1, t2, x2, y2 = rand_bbox_3d(batch.size(), lam)
for b in range(batch.size(0)):
batch[b, :, t1[b]:t2[b], x1[b]:x2[b], y1[b]:y2[b]] = batch[index[b], :, t1[b]:t2[b], x1[b]:x2[b], y1[b]:y2[b]]
lam = 1. - ((t2 - t1) * (x2 - x1) * (y2 - y1) / float((batch.size()[-1] * batch.size()[-2] * batch.size()[-3])))
elif batch.ndim == 4:
# 2D
x1, y1, x2, y2 = rand_bbox_2d(batch.size(), lam)
for b in range(batch.size(0)):
batch[b, :, x1[b]:x2[b], y1[b]:y2[b]] = batch[index[b], :, x1[b]:x2[b], y1[b]:y2[b]]
lam = 1. - ((x2 - x1) * (y2 - y1) / float((batch.size()[-1] * batch.size()[-2])))
return batch, index, lam
def cutmix_double_apply(batch, labels, alpha):
batch_size = batch.size(0)
lam = np.random.beta(alpha, alpha, batch_size)
lam = np.max((lam, 1.-lam), axis=0)
index = torch.randperm(batch_size)
# 2D - does not support 3D right now
x1, y1, x2, y2 = rand_bbox_2d(batch.size(), lam)
for b in range(batch.size(0)):
batch[b, :, x1[b]:x2[b], y1[b]:y2[b]] = batch[index[b], :, x1[b]:x2[b], y1[b]:y2[b]]
labels['seg'][b, :, x1[b]:x2[b], y1[b]:y2[b]] = labels['seg'][index[b], :, x1[b]:x2[b], y1[b]:y2[b]]
lam = 1. - ((x2 - x1) * (y2 - y1) / float((batch.size()[-1] * batch.size()[-2])))
class MixupBCELoss(nn.Module):
def __init__(self):
super().__init__()
def forward(self, y_pred, y_true):
if type(y_true) == dict:
# Training
y_true1 = y_true['y_true1']
y_true2 = y_true['y_true2']
lam = y_true['lam']
mix_loss1 = F.cross_entropy(y_pred, y_true1, reduction='none')
mix_loss2 = F.cross_entropy(y_pred, y_true2, reduction='none')
return (lam * mix_loss1 + (1. - lam) * mix_loss2).mean()
else:
# Validation
return F.binary_cross_entropy_with_logits(y_pred, y_true)
| [
"github@jph.am"
] | github@jph.am |
5e5ce0df1b1faf85f26ec4a9c54d6ac980b61e5a | 542f898adea1b36d627d4bf437731022f242d2dd | /projects/TridentNet/tridentnet/trident_backbone.py | 7789bd219b01d452e876ad2ad7f811502719465c | [
"Apache-2.0"
] | permissive | facebookresearch/detectron2 | 24bf508e374a98a5e5d1bd4cc96556d5914215f4 | 80307d2d5e06f06a8a677cc2653f23a4c56402ac | refs/heads/main | 2023-08-30T17:00:01.293772 | 2023-08-25T22:10:24 | 2023-08-25T22:10:24 | 206,660,580 | 27,469 | 8,047 | Apache-2.0 | 2023-09-13T09:25:57 | 2019-09-05T21:30:20 | Python | UTF-8 | Python | false | false | 7,846 | py | # Copyright (c) Facebook, Inc. and its affiliates.
import fvcore.nn.weight_init as weight_init
import torch
import torch.nn.functional as F
from detectron2.layers import Conv2d, FrozenBatchNorm2d, get_norm
from detectron2.modeling import BACKBONE_REGISTRY, ResNet, ResNetBlockBase
from detectron2.modeling.backbone.resnet import BasicStem, BottleneckBlock, DeformBottleneckBlock
from .trident_conv import TridentConv
__all__ = ["TridentBottleneckBlock", "make_trident_stage", "build_trident_resnet_backbone"]
class TridentBottleneckBlock(ResNetBlockBase):
def __init__(
self,
in_channels,
out_channels,
*,
bottleneck_channels,
stride=1,
num_groups=1,
norm="BN",
stride_in_1x1=False,
num_branch=3,
dilations=(1, 2, 3),
concat_output=False,
test_branch_idx=-1,
):
"""
Args:
num_branch (int): the number of branches in TridentNet.
dilations (tuple): the dilations of multiple branches in TridentNet.
concat_output (bool): if concatenate outputs of multiple branches in TridentNet.
Use 'True' for the last trident block.
"""
super().__init__(in_channels, out_channels, stride)
assert num_branch == len(dilations)
self.num_branch = num_branch
self.concat_output = concat_output
self.test_branch_idx = test_branch_idx
if in_channels != out_channels:
self.shortcut = Conv2d(
in_channels,
out_channels,
kernel_size=1,
stride=stride,
bias=False,
norm=get_norm(norm, out_channels),
)
else:
self.shortcut = None
stride_1x1, stride_3x3 = (stride, 1) if stride_in_1x1 else (1, stride)
self.conv1 = Conv2d(
in_channels,
bottleneck_channels,
kernel_size=1,
stride=stride_1x1,
bias=False,
norm=get_norm(norm, bottleneck_channels),
)
self.conv2 = TridentConv(
bottleneck_channels,
bottleneck_channels,
kernel_size=3,
stride=stride_3x3,
paddings=dilations,
bias=False,
groups=num_groups,
dilations=dilations,
num_branch=num_branch,
test_branch_idx=test_branch_idx,
norm=get_norm(norm, bottleneck_channels),
)
self.conv3 = Conv2d(
bottleneck_channels,
out_channels,
kernel_size=1,
bias=False,
norm=get_norm(norm, out_channels),
)
for layer in [self.conv1, self.conv2, self.conv3, self.shortcut]:
if layer is not None: # shortcut can be None
weight_init.c2_msra_fill(layer)
def forward(self, x):
num_branch = self.num_branch if self.training or self.test_branch_idx == -1 else 1
if not isinstance(x, list):
x = [x] * num_branch
out = [self.conv1(b) for b in x]
out = [F.relu_(b) for b in out]
out = self.conv2(out)
out = [F.relu_(b) for b in out]
out = [self.conv3(b) for b in out]
if self.shortcut is not None:
shortcut = [self.shortcut(b) for b in x]
else:
shortcut = x
out = [out_b + shortcut_b for out_b, shortcut_b in zip(out, shortcut)]
out = [F.relu_(b) for b in out]
if self.concat_output:
out = torch.cat(out)
return out
def make_trident_stage(block_class, num_blocks, **kwargs):
"""
Create a resnet stage by creating many blocks for TridentNet.
"""
concat_output = [False] * (num_blocks - 1) + [True]
kwargs["concat_output_per_block"] = concat_output
return ResNet.make_stage(block_class, num_blocks, **kwargs)
@BACKBONE_REGISTRY.register()
def build_trident_resnet_backbone(cfg, input_shape):
"""
Create a ResNet instance from config for TridentNet.
Returns:
ResNet: a :class:`ResNet` instance.
"""
# need registration of new blocks/stems?
norm = cfg.MODEL.RESNETS.NORM
stem = BasicStem(
in_channels=input_shape.channels,
out_channels=cfg.MODEL.RESNETS.STEM_OUT_CHANNELS,
norm=norm,
)
freeze_at = cfg.MODEL.BACKBONE.FREEZE_AT
if freeze_at >= 1:
for p in stem.parameters():
p.requires_grad = False
stem = FrozenBatchNorm2d.convert_frozen_batchnorm(stem)
# fmt: off
out_features = cfg.MODEL.RESNETS.OUT_FEATURES
depth = cfg.MODEL.RESNETS.DEPTH
num_groups = cfg.MODEL.RESNETS.NUM_GROUPS
width_per_group = cfg.MODEL.RESNETS.WIDTH_PER_GROUP
bottleneck_channels = num_groups * width_per_group
in_channels = cfg.MODEL.RESNETS.STEM_OUT_CHANNELS
out_channels = cfg.MODEL.RESNETS.RES2_OUT_CHANNELS
stride_in_1x1 = cfg.MODEL.RESNETS.STRIDE_IN_1X1
res5_dilation = cfg.MODEL.RESNETS.RES5_DILATION
deform_on_per_stage = cfg.MODEL.RESNETS.DEFORM_ON_PER_STAGE
deform_modulated = cfg.MODEL.RESNETS.DEFORM_MODULATED
deform_num_groups = cfg.MODEL.RESNETS.DEFORM_NUM_GROUPS
num_branch = cfg.MODEL.TRIDENT.NUM_BRANCH
branch_dilations = cfg.MODEL.TRIDENT.BRANCH_DILATIONS
trident_stage = cfg.MODEL.TRIDENT.TRIDENT_STAGE
test_branch_idx = cfg.MODEL.TRIDENT.TEST_BRANCH_IDX
# fmt: on
assert res5_dilation in {1, 2}, "res5_dilation cannot be {}.".format(res5_dilation)
num_blocks_per_stage = {50: [3, 4, 6, 3], 101: [3, 4, 23, 3], 152: [3, 8, 36, 3]}[depth]
stages = []
res_stage_idx = {"res2": 2, "res3": 3, "res4": 4, "res5": 5}
out_stage_idx = [res_stage_idx[f] for f in out_features]
trident_stage_idx = res_stage_idx[trident_stage]
max_stage_idx = max(out_stage_idx)
for idx, stage_idx in enumerate(range(2, max_stage_idx + 1)):
dilation = res5_dilation if stage_idx == 5 else 1
first_stride = 1 if idx == 0 or (stage_idx == 5 and dilation == 2) else 2
stage_kargs = {
"num_blocks": num_blocks_per_stage[idx],
"stride_per_block": [first_stride] + [1] * (num_blocks_per_stage[idx] - 1),
"in_channels": in_channels,
"bottleneck_channels": bottleneck_channels,
"out_channels": out_channels,
"num_groups": num_groups,
"norm": norm,
"stride_in_1x1": stride_in_1x1,
"dilation": dilation,
}
if stage_idx == trident_stage_idx:
assert not deform_on_per_stage[
idx
], "Not support deformable conv in Trident blocks yet."
stage_kargs["block_class"] = TridentBottleneckBlock
stage_kargs["num_branch"] = num_branch
stage_kargs["dilations"] = branch_dilations
stage_kargs["test_branch_idx"] = test_branch_idx
stage_kargs.pop("dilation")
elif deform_on_per_stage[idx]:
stage_kargs["block_class"] = DeformBottleneckBlock
stage_kargs["deform_modulated"] = deform_modulated
stage_kargs["deform_num_groups"] = deform_num_groups
else:
stage_kargs["block_class"] = BottleneckBlock
blocks = (
make_trident_stage(**stage_kargs)
if stage_idx == trident_stage_idx
else ResNet.make_stage(**stage_kargs)
)
in_channels = out_channels
out_channels *= 2
bottleneck_channels *= 2
if freeze_at >= stage_idx:
for block in blocks:
block.freeze()
stages.append(blocks)
return ResNet(stem, stages, out_features=out_features)
| [
"facebook-github-bot@users.noreply.github.com"
] | facebook-github-bot@users.noreply.github.com |
140384afde407034a54ba2db872c23687b2803b5 | f0d713996eb095bcdc701f3fab0a8110b8541cbb | /exeY2wDuEW4rFeYvL_18.py | df232bc3446da8ba44e538db19a12468c1434bda | [] | no_license | daniel-reich/turbo-robot | feda6c0523bb83ab8954b6d06302bfec5b16ebdf | a7a25c63097674c0a81675eed7e6b763785f1c41 | refs/heads/main | 2023-03-26T01:55:14.210264 | 2021-03-23T16:08:01 | 2021-03-23T16:08:01 | 350,773,815 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 854 | py | """
Create an ordered 2D list (matrix). A matrix is ordered if its (0, 0) element
is 1, its (0, 1) element is 2, and so on. Your function needs to create an a ×
b matrix. `a` is the first argument and `b` is the second.
### Examples
ordered_matrix(5, 5) ➞ [
[1, 2, 3, 4, 5],
[6, 7, 8, 9, 10],
[11, 12, 13, 14, 15],
[16, 17, 18, 19, 20],
[21, 22, 23, 24, 25]
]
ordered_matrix(1, 1) ➞ [[1]]
ordered_matrix(1, 5) ➞ [[1, 2, 3, 4, 5]]
### Notes
* `a` is the height of the matrix (y coordinate), and `b` is the width (x coordinate).
* `a` and `b` will always be positive, and the matrix will always be square shaped (in each row are the same amount of columns).
* `a` and `b` are integers.
"""
def ordered_matrix(a, b):
return [[b*i+j for j in range(1, b+1)] for i in range(a)]
| [
"daniel.reich@danielreichs-MacBook-Pro.local"
] | daniel.reich@danielreichs-MacBook-Pro.local |
55fbd18c40598e0709acb94e5d35fafdbae90167 | afc8d5a9b1c2dd476ea59a7211b455732806fdfd | /Configurations/ggH_SF/Full2017_v7/DYMVA_SYS/samples_recoil.py | dff83e4862c9124922705ab5e015b0c636e3c2f2 | [] | no_license | latinos/PlotsConfigurations | 6d88a5ad828dde4a7f45c68765081ed182fcda21 | 02417839021e2112e740607b0fb78e09b58c930f | refs/heads/master | 2023-08-18T20:39:31.954943 | 2023-08-18T09:23:34 | 2023-08-18T09:23:34 | 39,819,875 | 10 | 63 | null | 2023-08-10T14:08:04 | 2015-07-28T07:36:50 | Python | UTF-8 | Python | false | false | 16,488 | py | import inspect
configurations = os.path.realpath(inspect.getfile(inspect.currentframe())) # this file
configurations = os.path.dirname(configurations) # DYMVA_SYS
configurations = os.path.dirname(configurations) # Full2017_v7
configurations = os.path.dirname(configurations) # ggH_SF
configurations = os.path.dirname(configurations) # Configurations
from LatinoAnalysis.Tools.commonTools import getSampleFiles, getBaseW, addSampleWeight
def nanoGetSampleFiles(inputDir, sample):
try:
if _samples_noload:
return [sample]
except NameError:
pass
return getSampleFiles(inputDir, sample, True, 'nanoLatino_')
# samples
try:
len(samples)
except NameError:
import collections
samples = collections.OrderedDict()
################################################
################# SKIMS ########################
################################################
dataReco = 'Run2017_102X_nAODv7_Full2017v7'
fakeReco = dataReco
mcProduction = 'Fall2017_102X_nAODv7_Full2017v7'
mcSteps = 'MCl1loose2017v7__MCCorr2017v7__l2loose__l2tightOR2017v7{var}'
mcSteps_met = 'MCl1loose2017v7__MCCorr2017v7__l2loose__l2tightOR2017v7__recoilDY' #with MET recoil
fakeSteps = 'DATAl1loose2017v7__l2loose__fakeW'
dataSteps = 'DATAl1loose2017v7__l2loose__l2tightOR2017v7'
##############################################
###### Tree base directory for the site ######
##############################################
SITE=os.uname()[1]
if 'iihe' in SITE:
#treeBaseDir = '/pnfs/iihe/cms/store/user/xjanssen/HWW2015'
treeBaseDir = '/pnfs/iihe/cms/store/group/phys_higgs/cmshww/amassiro/HWWNano'
elif 'cern' in SITE:
treeBaseDir = '/eos/cms/store/group/phys_higgs/cmshww/amassiro/HWWNano'
def makeMCDirectory(var=''):
if var:
return os.path.join(treeBaseDir, mcProduction, mcSteps.format(var='__' + var))
else:
return os.path.join(treeBaseDir, mcProduction, mcSteps.format(var=''))
mcDirectory = makeMCDirectory()
metmcDirectory = os.path.join(treeBaseDir, mcProduction, mcSteps_met)
fakeDirectory = os.path.join(treeBaseDir, fakeReco, fakeSteps)
dataDirectory = os.path.join(treeBaseDir, dataReco, dataSteps)
################################################
############ DATA DECLARATION ##################
################################################
DataRun = [
['B','Run2017B-02Apr2020-v1'],
['C','Run2017C-02Apr2020-v1'],
['D','Run2017D-02Apr2020-v1'],
['E','Run2017E-02Apr2020-v1'],
['F','Run2017F-02Apr2020-v1']
]
DataSets = ['MuonEG','SingleMuon','SingleElectron','DoubleMuon', 'DoubleEG']
DataTrig = {
'MuonEG' : ' Trigger_ElMu' ,
'SingleMuon' : '!Trigger_ElMu && Trigger_sngMu' ,
'SingleElectron' : '!Trigger_ElMu && !Trigger_sngMu && Trigger_sngEl',
'DoubleMuon' : '!Trigger_ElMu && !Trigger_sngMu && !Trigger_sngEl && Trigger_dblMu',
'DoubleEG' : '!Trigger_ElMu && !Trigger_sngMu && !Trigger_sngEl && !Trigger_dblMu && Trigger_dblEl'
}
#########################################
############ MC COMMON ##################
#########################################
# SFweight does not include btag weights
mcCommonWeightNoMatch = 'XSWeight*SFweight*METFilter_MC'
mcCommonWeight = 'XSWeight*SFweight*PromptGenLepMatch2l*METFilter_MC'
###########################################
############# BACKGROUNDS ###############
###########################################
###### DY #######
useDYtt = False
useDYHT = True
# ptllDYW_NLO = '(((0.623108 + 0.0722934*gen_ptll - 0.00364918*gen_ptll*gen_ptll + 6.97227e-05*gen_ptll*gen_ptll*gen_ptll - 4.52903e-07*gen_ptll*gen_ptll*gen_ptll*gen_ptll)*(gen_ptll<45)*(gen_ptll>0) + 1*(gen_ptll>=45))*(abs(gen_mll-90)<3) + (abs(gen_mll-90)>3))'
# ptllDYW_LO = '((0.632927+0.0456956*gen_ptll-0.00154485*gen_ptll*gen_ptll+2.64397e-05*gen_ptll*gen_ptll*gen_ptll-2.19374e-07*gen_ptll*gen_ptll*gen_ptll*gen_ptll+6.99751e-10*gen_ptll*gen_ptll*gen_ptll*gen_ptll*gen_ptll)*(gen_ptll>0)*(gen_ptll<100)+(1.41713-0.00165342*gen_ptll)*(gen_ptll>=100)*(gen_ptll<300)+1*(gen_ptll>=300))'
if useDYtt:
files = nanoGetSampleFiles(mcDirectory, 'DYJetsToTT_MuEle_M-50') + \
nanoGetSampleFiles(mcDirectory, 'DYJetsToLL_M-10to50-LO_ext1')
samples['DY'] = {
'name': files,
'weight': mcCommonWeight + "*( !(Sum$(PhotonGen_isPrompt==1 && PhotonGen_pt>15 && abs(PhotonGen_eta)<2.6) > 0 &&\
Sum$(LeptonGen_isPrompt==1 && LeptonGen_pt>15)>=2) )",
'FilesPerJob': 5,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
}
addSampleWeight(samples,'DY','DYJetsToTT_MuEle_M-50','DY_NLO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-10to50-LO_ext1','DY_LO_pTllrw')
else:
files = nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50') + \
nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-10to50-LO_ext1')
samples['DY'] = {
'name': files,
'weight': mcCommonWeight + "*( !(Sum$(PhotonGen_isPrompt==1 && PhotonGen_pt>15 && abs(PhotonGen_eta)<2.6) > 0 &&\
Sum$(LeptonGen_isPrompt==1 && LeptonGen_pt>15)>=2) )",
'FilesPerJob': 8,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
}
# Add DY HT Samples
if useDYHT :
samples['DY']['name'] += nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-4to50_HT-100to200_ext1') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-4to50_HT-200to400_newpmx') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-4to50_HT-400to600') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-4to50_HT-600toInf') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-100to200') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-200to400') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-400to600_ext1') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-600to800') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-800to1200') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-1200to2500') \
+ nanoGetSampleFiles(metmcDirectory, 'DYJetsToLL_M-50_HT-2500toInf')
addSampleWeight(samples,'DY','DYJetsToLL_M-50','DY_NLO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-10to50-LO_ext1','DY_LO_pTllrw')
if useDYHT :
# Remove high HT from inclusive samples
addSampleWeight(samples,'DY','DYJetsToLL_M-50' , 'LHE_HT<100.0')
addSampleWeight(samples,'DY','DYJetsToLL_M-10to50-LO_ext1', 'LHE_HT<100.0')
# pt_ll weight
addSampleWeight(samples,'DY','DYJetsToLL_M-4to50_HT-100to200_ext1' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-4to50_HT-200to400_newpmx' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-4to50_HT-400to600' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-4to50_HT-600toInf' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-100to200' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-200to400' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-400to600_ext1' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-600to800' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-800to1200' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-1200to2500' ,'DY_LO_pTllrw')
addSampleWeight(samples,'DY','DYJetsToLL_M-50_HT-2500toInf' ,'DY_LO_pTllrw')
###### Top #######
files = nanoGetSampleFiles(mcDirectory, 'TTTo2L2Nu') + \
nanoGetSampleFiles(mcDirectory, 'ST_s-channel') + \
nanoGetSampleFiles(mcDirectory, 'ST_t-channel_antitop') + \
nanoGetSampleFiles(mcDirectory, 'ST_t-channel_top') + \
nanoGetSampleFiles(mcDirectory, 'ST_tW_antitop') + \
nanoGetSampleFiles(mcDirectory, 'ST_tW_top')
samples['top'] = {
'name': files,
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 1,
}
addSampleWeight(samples,'top','TTTo2L2Nu','Top_pTrw')
###### WW ########
samples['WW'] = {
'name': nanoGetSampleFiles(mcDirectory, 'WWTo2L2Nu'),
'weight': mcCommonWeight + '*nllW',
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 1
}
samples['WWewk'] = {
'name': nanoGetSampleFiles(mcDirectory, 'WpWmJJ_EWK_noTop'),
'weight': mcCommonWeight + '*(Sum$(abs(GenPart_pdgId)==6 || GenPart_pdgId==25)==0)', #filter tops and Higgs
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
# k-factor 1.4 already taken into account in XSWeight
files = nanoGetSampleFiles(mcDirectory, 'GluGluToWWToENEN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToENMN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToENTN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToMNEN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToMNMN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToMNTN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToTNEN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToTNMN') + \
nanoGetSampleFiles(mcDirectory, 'GluGluToWWToTNTN')
samples['ggWW'] = {
'name': files,
'weight': mcCommonWeight + '*1.53/1.4', # updating k-factor
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 10
}
######## Vg ########
files = nanoGetSampleFiles(mcDirectory, 'Wg_MADGRAPHMLM') + \
nanoGetSampleFiles(mcDirectory, 'ZGToLLG')
samples['Vg'] = {
'name': files,
'weight': mcCommonWeightNoMatch + '*!(Gen_ZGstar_mass > 0)',
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 10
}
######## VgS ########
files = nanoGetSampleFiles(mcDirectory, 'Wg_MADGRAPHMLM') + \
nanoGetSampleFiles(mcDirectory, 'ZGToLLG') + \
nanoGetSampleFiles(mcDirectory, 'WZTo3LNu_mllmin01')
samples['VgS'] = {
'name': files,
'weight': mcCommonWeight + ' * (gstarLow * 0.94 + gstarHigh * 1.14)',
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 15,
'subsamples': {
'L': 'gstarLow',
'H': 'gstarHigh'
}
}
addSampleWeight(samples, 'VgS', 'Wg_MADGRAPHMLM', '(Gen_ZGstar_mass > 0 && Gen_ZGstar_mass < 0.1)')
addSampleWeight(samples, 'VgS', 'ZGToLLG', '(Gen_ZGstar_mass > 0)')
addSampleWeight(samples, 'VgS', 'WZTo3LNu_mllmin01', '(Gen_ZGstar_mass > 0.1)')
############ VZ ############
files = nanoGetSampleFiles(mcDirectory, 'ZZTo2L2Nu') + \
nanoGetSampleFiles(mcDirectory, 'ZZTo2L2Q') + \
nanoGetSampleFiles(mcDirectory, 'ZZTo4L') + \
nanoGetSampleFiles(mcDirectory, 'WZTo2L2Q')
samples['VZ'] = {
'name': files,
'weight': mcCommonWeight + '*1.11',
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
########## VVV #########
files = nanoGetSampleFiles(mcDirectory, 'ZZZ') + \
nanoGetSampleFiles(mcDirectory, 'WZZ') + \
nanoGetSampleFiles(mcDirectory, 'WWZ') + \
nanoGetSampleFiles(mcDirectory, 'WWW')
#+ nanoGetSampleFiles(mcDirectory, 'WWG'), #should this be included? or is it already taken into account in the WW sample?
samples['VVV'] = {
'name': files,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'weight': mcCommonWeight
}
###########################################
############# SIGNALS ##################
###########################################
signals = []
#### ggH -> WW
samples['ggH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'GluGluHToWWTo2L2Nu_Powheg_M125')+nanoGetSampleFiles(mcDirectory, 'GGHjjToWWTo2L2Nu_minloHJJ_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 4,
}
addSampleWeight(samples, 'ggH_hww', 'GluGluHToWWTo2L2Nu_Powheg_M125', '(HTXS_stage1_1_cat_pTjet30GeV<107)*Weight2MINLO*1093.8199/1073.9094') #only non GE2J categories with the weight to NNLOPS and renormalize integral
addSampleWeight(samples, 'ggH_hww', 'GGHjjToWWTo2L2Nu_minloHJJ_M125', '(HTXS_stage1_1_cat_pTjet30GeV>106)*1093.8199/1073.9094')
signals.append('ggH_hww')
############ VBF H->WW ############
samples['qqH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'VBFHToWWTo2L2Nu_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 3
}
signals.append('qqH_hww')
############# ZH H->WW ############
samples['ZH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'HZJ_HToWWTo2L2Nu_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 1
}
signals.append('ZH_hww')
samples['ggZH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'GluGluZH_HToWWTo2L2Nu_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
signals.append('ggZH_hww')
############ WH H->WW ############
samples['WH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'HWplusJ_HToWW_M125') + nanoGetSampleFiles(mcDirectory, 'HWminusJ_HToWW_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
signals.append('WH_hww')
############ ttH ############
samples['ttH_hww'] = {
'name': nanoGetSampleFiles(mcDirectory, 'ttHToNonbb_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 1
}
signals.append('ttH_hww')
############ H->TauTau ############
samples['ggH_htt'] = {
'name': nanoGetSampleFiles(mcDirectory, 'GluGluHToTauTau_M125_ext1'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 1
}
#signals.append('ggH_htt')
samples['qqH_htt'] = {
'name': nanoGetSampleFiles(mcDirectory, 'VBFHToTauTau_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
#signals.append('qqH_htt')
samples['ZH_htt'] = {
'name': nanoGetSampleFiles(mcDirectory, 'HZJ_HToTauTau_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
#signals.append('ZH_htt')
samples['WH_htt'] = {
'name': nanoGetSampleFiles(mcDirectory, 'HWplusJ_HToTauTau_M125') + nanoGetSampleFiles(mcDirectory, 'HWminusJ_HToTauTau_M125'),
'weight': mcCommonWeight,
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 2
}
#signals.append('WH_htt')
###########################################
################## FAKE ###################
###########################################
samples['Fake'] = {
'name': [],
'weight': 'METFilter_DATA*fakeW',
'weights': [],
'isData': ['all'],
'suppressNegative' :['all'],
'suppressNegativeNuisances' :['all'],
'FilesPerJob': 30
}
for _, sd in DataRun:
for pd in DataSets:
files = nanoGetSampleFiles(fakeDirectory, pd + '_' + sd)
samples['Fake']['name'].extend(files)
samples['Fake']['weights'].extend([DataTrig[pd]] * len(files))
samples['Fake']['subsamples'] = {
'ee': 'abs(Lepton_pdgId[0]) == 11 && abs(Lepton_pdgId[1]) == 11',
'mm': 'abs(Lepton_pdgId[0]) == 13 && abs(Lepton_pdgId[1]) == 13',
'df': '(Lepton_pdgId[0]*Lepton_pdgId[1] == -11*13)'
}
###########################################
################## DATA ###################
###########################################
samples['DATA'] = {
'name': [],
'weight': 'METFilter_DATA*LepWPCut',
'weights': [],
'isData': ['all'],
'FilesPerJob': 40
}
for _, sd in DataRun:
for pd in DataSets:
files = nanoGetSampleFiles(dataDirectory, pd + '_' + sd)
samples['DATA']['name'].extend(files)
samples['DATA']['weights'].extend([DataTrig[pd]] * len(files))
| [
"calderon@cern.ch"
] | calderon@cern.ch |
96c01056fd75ce8b7db7ce7ccb777c86f47c27d1 | fcf73ee47db9df86631753a7faaa8c31ba7def16 | /hawkentracker/alembic/versions/55a540177a3_explict_opt_out_status.py | 2936d1ced640074545157dcbcd57a9882a3bc569 | [
"MIT"
] | permissive | ashfire908/hawken-tracker | 4079ea3df9603bc4216b5a1fdf95bf82fdbe1987 | 4db02a63237100da43f3b8f44a6ebd862cf5d8cc | refs/heads/master | 2021-03-22T04:55:57.822803 | 2018-01-02T20:48:40 | 2018-01-02T20:48:40 | 42,843,654 | 5 | 1 | null | null | null | null | UTF-8 | Python | false | false | 971 | py | """Explict opt out status
Revision ID: 55a540177a3
Revises: 276a9c91812
Create Date: 2015-03-30 13:21:10.374065
"""
# revision identifiers, used by Alembic.
revision = "55a540177a3"
down_revision = "276a9c91812"
from alembic import op
import sqlalchemy as sa
def upgrade():
# Allow for nullable opt out
op.alter_column("players", "opt_out", nullable=True)
# Convert false opt out to null
players = sa.sql.table("players",
sa.Column("opt_out", sa.Boolean, nullable=True)
)
op.execute(
players.update().where(players.c.opt_out == False).values({"opt_out": None})
)
def downgrade():
# Convert null opt out to false
players = sa.sql.table("players",
sa.Column("opt_out", sa.Boolean, nullable=True)
)
op.execute(
players.update().where(players.c.opt_out == None).values({"opt_out": False})
)
# Disallow for nullable opt out
op.alter_column("players", "opt_out", nullable=False)
| [
"andrew.hampe@gmail.com"
] | andrew.hampe@gmail.com |
8b54cfc2b902f28ab710f11c95b3372f1a69a143 | bbd6af23c1c56bf090ddaa46ae639d128f91c5eb | /ourproj/ourproj/spiders/ourspider.py | 9965ea9c65a6dfbd60a27ecba67c4e360d3d92d2 | [
"BSD-3-Clause"
] | permissive | nichgaun/scrapy | 8f75bdfc713ec845e6f7d5666f28f4756fa1f819 | 493918742d8359583bdba950ad1aad43123fc394 | refs/heads/master | 2020-03-09T07:13:56.572918 | 2018-04-16T22:37:38 | 2018-04-16T22:37:38 | 128,659,805 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 818 | py | # -*- coding: utf-8 -*-
import scrapy
from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.contrib.linkextractors.lxmlhtml import LxmlLinkExtractor
class OurspiderSpider(CrawlSpider):
name = 'ourspider'
allowed_domains = ['https://en.wikipedia.org']
start_urls = ['https://en.wikipedia.org/wiki/Lewis_Tappan_Barney']
rules = (Rule(LxmlLinkExtractor(allow=()), callback='parse_obj', follow=True),)
#def start_requests(self):
# yield scrapy.Request('https://en.wikipedia.org', meta={'bindaddress': ('1234:5678:111::0a', 0)})
# def parse(self, response):
# pass
def parse_obj(self,response):
for link in LxmlLinkExtractor(allow=()).extract_links(response):
print(link);
# item = someItem()
# item['url'] = link.url
| [
"paulctroost@gmail.com"
] | paulctroost@gmail.com |
6918cee393ce45548af443bf75a05328fc982ad2 | 96c3ca6ad097a024c6d7eabb79d715aca0d1ab66 | /Auth/migrations/0003_auto_20210626_1422.py | 1027129046e4682023e1972313d566602b210a8c | [] | no_license | aimperatori/RFID-API2 | 937cd4e29787688da962f21d95bae67fcad9ca07 | addf00fd1c9efcbea62910a4126e678e00d4cbdd | refs/heads/main | 2023-06-04T15:54:53.661517 | 2021-06-26T16:59:16 | 2021-06-26T16:59:16 | 380,553,419 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 540 | py | # Generated by Django 3.2.4 on 2021-06-26 14:22
from django.db import migrations, models
class Migration(migrations.Migration):
dependencies = [
('Auth', '0002_alter_auth_created_datetime'),
]
operations = [
migrations.RemoveField(
model_name='auth',
name='id',
),
migrations.AlterField(
model_name='auth',
name='rdifCode',
field=models.CharField(editable=False, max_length=16, primary_key=True, serialize=False),
),
]
| [
"aimperatori@ucs.br"
] | aimperatori@ucs.br |
3686504c9548d45692d16c538ab9fac95904d340 | 23cb0778282544457b76cc5d90501a3e142b89b9 | /simplified_scrapy/core/xml_helper.py | bc0c20ac8fd6f940b2f2b1ad7363f02924c72cb3 | [
"Apache-2.0",
"LicenseRef-scancode-warranty-disclaimer"
] | permissive | ra2003/simplified-scrapy | bcc4bba8e6a07cceebe0017a5538cb2909b6a6b1 | 4a9b1913b716685e57baa5e8032bf4367060fc98 | refs/heads/master | 2022-04-17T07:11:12.087373 | 2020-04-03T00:10:28 | 2020-04-03T00:10:28 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,529 | py | #!/usr/bin/python
#coding=utf-8
try:
import xml.etree.cElementTree as ET
except ImportError:
import xml.etree.ElementTree as ET
import sys,re
from simplified_scrapy.core.utils import printInfo
from simplified_scrapy.core.dictex import Dict
class XmlDictConfig(Dict):
def __init__(self, parent_element):
if parent_element.items():
self.update(Dict(parent_element.items()))
flag = False
for element in parent_element:
flag = True
if(not self.get(element.tag)):
self.update({element.tag: []})
dic = self.getDic(element)
self[element.tag].append(dic)
count = len(element)
if(count>0):
self.ele2arr(dic,element)
if(not flag):
self.update({'tag':parent_element.tag})
def convert2Dic(html):
try:
tag=''
if(html.find('</')<0 and html.find('/>')<0):
start = html.find('<')
end = html.find(' ',start+1)
tag = '</'+html[start+1:end]+'>'
tree = ET.XML(html+tag)
return XmlDictConfig(tree)
except Exception as err:
try:
start = html.find('<')
end = html.find('>')
html = html[start+1:end].strip('/').strip()
html = re.sub('(\\s| )+', ' ', html, 0)
html = re.sub('(\')+', '"', html, 0)
html = re.sub('(=\s*")+', '="', html, 0)
lstC = []#list(html)
N=len(html)
i=0
first = False
flag = False
while i<N:
if html[i]=='"':
lstC.append(html[i])
first = not first
elif not first and html[i]=='=' and html[i+1]!='"':
lstC.append(html[i])
lstC.append('"')
flag=True
elif not first and flag and html[i]==' ':
flag=False
lstC.append('"')
lstC.append(html[i])
else:
lstC.append(html[i])
i+=1
html = ''.join(lstC)
paras = html.split('"')
dic = Dict()
lastP=None
first = True
for para in paras:
if(first):
first=False
tmp=para.split()
dic['tag']=tmp[0]
if(len(tmp)>1):
lastP=tmp[1].strip().strip('=').strip()
continue
if(lastP):
if(not dic[lastP]):
dic[lastP]=para
else:
dic[lastP]+=' '
dic[lastP]+=para
lastP=None
elif para:
if(para.find('=')>0):
lastP=para.strip().strip('=').strip()
else:
dic[para]=''
return dic
except Exception as err:
printInfo(err)
return None
| [
"3095069599@qq.com"
] | 3095069599@qq.com |
ea2c22d2bcc968840f2546a7797fd481f4baee63 | ccbfc7818c0b75929a1dfae41dc061d5e0b78519 | /aliyun-openapi-python-sdk-master/aliyun-python-sdk-polardb/aliyunsdkpolardb/request/v20170801/ModifyAccountPasswordRequest.py | f043c7559542a8e3d3c3580f0a4b31e7c654201e | [
"Apache-2.0"
] | permissive | P79N6A/dysms_python | 44b634ffb2856b81d5f79f65889bfd5232a9b546 | f44877b35817e103eed469a637813efffa1be3e4 | refs/heads/master | 2020-04-28T15:25:00.368913 | 2019-03-13T07:52:34 | 2019-03-13T07:52:34 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 2,300 | py | # Licensed to the Apache Software Foundation (ASF) under one
# or more contributor license agreements. See the NOTICE file
# distributed with this work for additional information
# regarding copyright ownership. The ASF licenses this file
# to you under the Apache License, Version 2.0 (the
# "License"); you may not use this file except in compliance
# with the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
#
#
# Unless required by applicable law or agreed to in writing,
# software distributed under the License is distributed on an
# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
# KIND, either express or implied. See the License for the
# specific language governing permissions and limitations
# under the License.
from aliyunsdkcore.request import RpcRequest
class ModifyAccountPasswordRequest(RpcRequest):
def __init__(self):
RpcRequest.__init__(self, 'polardb', '2017-08-01', 'ModifyAccountPassword','polardb')
def get_ResourceOwnerId(self):
return self.get_query_params().get('ResourceOwnerId')
def set_ResourceOwnerId(self,ResourceOwnerId):
self.add_query_param('ResourceOwnerId',ResourceOwnerId)
def get_AccountName(self):
return self.get_query_params().get('AccountName')
def set_AccountName(self,AccountName):
self.add_query_param('AccountName',AccountName)
def get_NewAccountPassword(self):
return self.get_query_params().get('NewAccountPassword')
def set_NewAccountPassword(self,NewAccountPassword):
self.add_query_param('NewAccountPassword',NewAccountPassword)
def get_ResourceOwnerAccount(self):
return self.get_query_params().get('ResourceOwnerAccount')
def set_ResourceOwnerAccount(self,ResourceOwnerAccount):
self.add_query_param('ResourceOwnerAccount',ResourceOwnerAccount)
def get_DBClusterId(self):
return self.get_query_params().get('DBClusterId')
def set_DBClusterId(self,DBClusterId):
self.add_query_param('DBClusterId',DBClusterId)
def get_OwnerAccount(self):
return self.get_query_params().get('OwnerAccount')
def set_OwnerAccount(self,OwnerAccount):
self.add_query_param('OwnerAccount',OwnerAccount)
def get_OwnerId(self):
return self.get_query_params().get('OwnerId')
def set_OwnerId(self,OwnerId):
self.add_query_param('OwnerId',OwnerId) | [
"1478458905@qq.com"
] | 1478458905@qq.com |
99d16f620ac24b74834e13c63e09b6196c038fb0 | 7f4fb112bc9ab2b90f5f2248f43285ce9ac2e0a0 | /src/igem/neutronics/air/bare/borosilicate-glass-backfill/0wt/plot_all.in.one_cask.thickness_dose.rate_t4045_plug.py | c763b91a79e02074300d90606bbccfa7b9fb3d2b | [] | no_license | TheDoctorRAB/plot | dd3b5134c91c8fa7032fcc077c5427b26a80e49d | ed6746d511222c03e79f93548fe3ecd4286bf7b1 | refs/heads/master | 2021-07-11T10:21:19.347531 | 2020-07-16T17:13:15 | 2020-07-16T17:13:15 | 20,462,189 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 6,119 | py | ########################################################################
# R.A.Borrelli
# @TheDoctorRAB
# rev.11.March.2015
########################################################################
#
# Plot routine
# All in one file, with no separate control input, lib files
# Plot data is contained in a separate data file, read on command line
# Set up for a secondary y axis if needed
#
########################################################################
#
#
#
#######
#
# imports
#
# plot
#
import numpy
import matplotlib
import matplotlib.pyplot as plot
from matplotlib.ticker import MultipleLocator
#
#######
#
# command line
#
from sys import argv
script,plot_datafile=argv #column 0 is the x values then odd columns contain dose/flux
#
#######
#
# screen resolution
#
import Tkinter
root=Tkinter.Tk()
#
########################################################################
#
#
#
#######
#
# screen resolution
#
###
#
# pixels
#
width=root.winfo_screenwidth()
height=root.winfo_screenheight()
#
###
#
# mm
#
width_mm=root.winfo_screenmmwidth()
height_mm=root.winfo_screenmmheight()
#
###
#
# in
#
width_in=width_mm/25.4
height_in=height_mm/25.4
#
###
#
# dpi
#
width_dpi=width/width_in
height_dpi=height/height_in
#
dpi_values=(96,120,144,168,192)
current_dpi=width_dpi
minimum=1000
#
for dval in dpi_values:
difference=abs(dval-width_dpi)
if difference<minimum:
minimum=difference
current_dpi=dval
#
#######
#
# output to screen
#
print('width: %i px, height: %i px'%(width,height))
print('width: %i mm, height: %i mm'%(width_mm,height_mm))
print('width: %0.f in, height: %0.f in'%(width_in,height_in))
print('width: %0.f dpi, height: %0.f dpi'%(width_dpi,height_dpi))
print('size is %0.f %0.f'%(width,height))
print('current DPI is %0.f' % (current_dpi))
#
#######
#
# open the plot data file(s)
# add plot_dataN for each plot_datafileN
#
plot_data=numpy.loadtxt(plot_datafile,dtype=float)
#
#######
#
# graph parameters
#
###
#
# font sizes
#
matplotlib.rcParams.update({'font.size': 48}) #axis numbers
#
title_fontsize=54 #plot title
axis_fontsize=48 #axis labels
annotate_fontsize=48 #annotation
#
###
#
# set up for two y axis
#
fig,left_axis=plot.subplots()
# right_axis=left_axis.twinx()
#
###
#
# plot text
#
title='Dose rate - Bottom plate'
xtitle='Wall thickness [cm]'
ytitle='Dose rate [$\mu$Sv/h]'
#
###
#
# legend
# add linecolorN for each plot_dataN
# add curve_textN for each plot_dataN
#
line_color0='blue' #color
line_color1='orange' #color
line_color2='red' #color
line_color3='green' #color
line_color4='cyan' #color
#
curve_text0='10 wt% $B_4C$' #legend text
curve_text1='30 wt% $B_4C$' #legend text
curve_text2='50 wt% $B_4C$' #legend text
curve_text3='70 wt% $B_4C$' #legend text
curve_text4='90 wt% $B_4C$' #legend text
#
legend_location='lower left' #location of legend on grid
legend_font=42
#
###
#
# annotate
# position of the annotation dependent on axis domain and range
#
annotate_title='T-4045'
annotate_x=23
annotate_y=10000
#
annotate_title2='Air-Glass backfill'
annotate_x2=23
annotate_y2=7000
#
annotate_title3='0 wt% $^{10}B$'
annotate_x3=23
annotate_y3=3000
#
###
#
# axis domain and range
#
xmin=1
xmax=31
#
ymin=1
ymax=15000
#
###
#
# axis ticks
#
xmajortick=5
ymajortick=5000
#
xminortick=1
yminortick=1000
#
###
#
# grid linewidth
#
major_grid_linewidth=2.5
minor_grid_linewidth=2.1
#
major_grid_tick_length=7
minor_grid_tick_length=5
#
###
#
# curve linewidth
#
curve_linewidth=4.0
#
#######
#
# set plot diagnostics
#
###
#
# titles
#
plot.title(title,fontsize=title_fontsize)
left_axis.set_xlabel(xtitle,fontsize=axis_fontsize)
left_axis.set_ylabel(ytitle,fontsize=axis_fontsize)
# right_axis.set_ylabel()
#
###
#
# grid
#
left_axis.grid(which='major',axis='both',linewidth=major_grid_linewidth)
left_axis.grid(which='minor',axis='both',linewidth=minor_grid_linewidth)
#
left_axis.tick_params(axis='both',which='major',direction='inout',length=major_grid_tick_length)
left_axis.tick_params(axis='both',which='minor',direction='inout',length=minor_grid_tick_length)
#
###
#
# axis domain and range
#
plot.xlim(xmin,xmax)
left_axis.axis(ymin=ymin,ymax=ymax)
###
#
# axis ticks
#
left_axis.xaxis.set_major_locator(MultipleLocator(xmajortick))
left_axis.xaxis.set_minor_locator(MultipleLocator(xminortick))
left_axis.yaxis.set_major_locator(MultipleLocator(ymajortick))
left_axis.yaxis.set_minor_locator(MultipleLocator(yminortick))
#
###
#
# log scale option
# xmin,ymin !=0 for log scale
#
#left_axis.set_xscale('log')
left_axis.set_yscale('log')
#
###
#
# annotation
# comment out if not needed
#
left_axis.annotate(annotate_title,xy=(annotate_x,annotate_y),xytext=(annotate_x,annotate_y),fontsize=annotate_fontsize)
left_axis.annotate(annotate_title2,xy=(annotate_x2,annotate_y2),xytext=(annotate_x2,annotate_y2),fontsize=annotate_fontsize)
left_axis.annotate(annotate_title3,xy=(annotate_x3,annotate_y3),xytext=(annotate_x3,annotate_y3),fontsize=annotate_fontsize)
#
#######
#
# plot data
#
left_axis.plot(plot_data[:,0],plot_data[:,1],marker='o',color=line_color0,label=curve_text0,linewidth=curve_linewidth,markersize=20)
left_axis.plot(plot_data[:,0],plot_data[:,3],marker='o',color=line_color1,label=curve_text1,linewidth=curve_linewidth,markersize=20)
left_axis.plot(plot_data[:,0],plot_data[:,5],marker='o',color=line_color2,label=curve_text2,linewidth=curve_linewidth,markersize=20)
left_axis.plot(plot_data[:,0],plot_data[:,7],marker='o',color=line_color3,label=curve_text3,linewidth=curve_linewidth,markersize=20)
left_axis.plot(plot_data[:,0],plot_data[:,9],marker='o',color=line_color4,label=curve_text4,linewidth=curve_linewidth,markersize=20)
left_axis.legend(loc=legend_location,fontsize=legend_font) #legend needs to be after all the plot data
plot.get_current_fig_manager().resize(width,height)
plot.gcf().set_size_inches((0.01*width),(0.01*height))
#
#######
#
# save
#
plot.savefig(title,dpi=current_dpi)
#
#######
#
# plot to screen
#
# plot.show()
#
########################################################################
#
# EOF
#
########################################################################
| [
"borrelli@localhost.localdomain"
] | borrelli@localhost.localdomain |
5147440f426f86123ef2afa942129888aa4c2655 | 50ee312e98af4531330b66d396013cafefae87e7 | /softlearning/environments/gym/mujoco/walker2d_env.py | 3ed2d86cc402c67b34f62fd22a2c1f2eec171a56 | [
"MIT",
"LicenseRef-scancode-generic-cla"
] | permissive | richardrl/softlearning | 5f38fd136f7bcb95797818ae5261d4058bcfbb13 | 125ee6ee137145947703018e9980d064c94b1666 | refs/heads/master | 2020-04-21T06:27:46.778091 | 2019-02-02T20:36:38 | 2019-02-02T20:36:38 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 3,899 | py | import numpy as np
from gym.envs.mujoco import mujoco_env
from gym import utils
DEFAULT_CAMERA_CONFIG = {
'trackbodyid': 2,
'distance': 4.0,
'lookat': (None, None, 1.15),
'elevation': -20.0,
}
class Walker2dEnv(mujoco_env.MujocoEnv, utils.EzPickle):
def __init__(self,
forward_reward_weight=1.0,
ctrl_cost_weight=1e-3,
healthy_reward=1.0,
terminate_when_unhealthy=True,
healthy_z_range=(0.8, 2.0),
healthy_angle_range=(-1.0, 1.0),
reset_noise_scale=5e-3,
exclude_current_positions_from_observation=True):
utils.EzPickle.__init__(**locals())
self._forward_reward_weight = forward_reward_weight
self._ctrl_cost_weight = ctrl_cost_weight
self._healthy_reward = healthy_reward
self._terminate_when_unhealthy = terminate_when_unhealthy
self._healthy_z_range = healthy_z_range
self._healthy_angle_range = healthy_angle_range
self._reset_noise_scale = reset_noise_scale
self._exclude_current_positions_from_observation = (
exclude_current_positions_from_observation)
mujoco_env.MujocoEnv.__init__(self, "walker2d.xml", 4)
@property
def healthy_reward(self):
return float(
self.is_healthy
or self._terminate_when_unhealthy
) * self._healthy_reward
def control_cost(self, action):
control_cost = self._ctrl_cost_weight * np.sum(np.square(action))
return control_cost
@property
def is_healthy(self):
z, angle = self.sim.data.qpos[1:3]
min_z, max_z = self._healthy_z_range
min_angle, max_angle = self._healthy_angle_range
healthy_z = min_z < z < max_z
healthy_angle = min_angle < angle < max_angle
is_healthy = healthy_z and healthy_angle
return is_healthy
@property
def done(self):
done = (not self.is_healthy
if self._terminate_when_unhealthy
else False)
return done
def _get_obs(self):
position = self.sim.data.qpos.flat.copy()
velocity = np.clip(
self.sim.data.qvel.flat.copy(), -10, 10)
if self._exclude_current_positions_from_observation:
position = position[1:]
observation = np.concatenate((position, velocity)).ravel()
return observation
def step(self, action):
x_position_before = self.sim.data.qpos[0]
self.do_simulation(action, self.frame_skip)
x_position_after = self.sim.data.qpos[0]
x_velocity = ((x_position_after - x_position_before)
/ self.dt)
ctrl_cost = self.control_cost(action)
forward_reward = self._forward_reward_weight * x_velocity
healthy_reward = self.healthy_reward
rewards = forward_reward + healthy_reward
costs = ctrl_cost
observation = self._get_obs()
reward = rewards - costs
done = self.done
info = {
'x_position': x_position_after,
'x_velocity': x_velocity,
}
return observation, reward, done, info
def reset_model(self):
noise_low = -self._reset_noise_scale
noise_high = self._reset_noise_scale
qpos = self.init_qpos + self.np_random.uniform(
low=noise_low, high=noise_high, size=self.model.nq)
qvel = self.init_qvel + self.np_random.uniform(
low=noise_low, high=noise_high, size=self.model.nv)
self.set_state(qpos, qvel)
observation = self._get_obs()
return observation
def viewer_setup(self):
self.viewer.cam.trackbodyid = 2
self.viewer.cam.distance = self.model.stat.extent * 0.5
self.viewer.cam.lookat[2] = 1.15
self.viewer.cam.elevation = -20
| [
"hartikainen@berkeley.edu"
] | hartikainen@berkeley.edu |
4539fad165238fc206bb83ea657ad5f85e84cb86 | 642911284dff300708f9a777c9792eae2bd4c256 | /orgCodes/test.py | 9e5bab8c012a36aa0276473a0a3457f3b0981563 | [] | no_license | keyman9848/bncgit | 3b6aea4bb3bd229e0ae96d15becb9170be134b41 | 915972fe2012024b6e87aaa48a5350dfe815e6e3 | refs/heads/master | 2021-01-25T16:53:45.188797 | 2017-08-24T08:23:54 | 2017-08-24T08:23:54 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 4,007 | py | #-*_coding:utf8-*-
from multiprocessing.dummy import Pool as ThreadPool
import re,requests
import urllib.request
class spider(object):
def get_source(self,url):
hds={
'Connection': 'Keep-Alive',
'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,*/*;q=0.8',
'Accept-Language': 'zh-CN,zh;q=0.8,en;q=0.6',
'User-Agent': 'Googlebot/2.1 (+http://www.googlebot.com/bot.html)',
'Host': 'www.tianyancha.com',
'Referer': 'http://antirobot.tianyancha.com/captcha/verify?return_url=http://www.tianyancha.com/search/%E6%B1%9F%E8%A5%BF%20%20%20%E4%BA%BA%E5%8A%9B%E8%B5%84%E6%BA%90/11'
}
req = urllib.request.Request(url, data=None, headers=hds)
response = urllib.request.urlopen(req)
return response.read()
def get_companyurl(self,source):
companyurl=re.findall('<a href="(.*)" ng-click',source)
companyurl_group=[]
for j in companyurl:
every_companyurl='http://www.tianyancha.com' + j
companyurl_group.append(every_companyurl)
return companyurl_group
def get_companyinfo(self,source):
try:
info={}
company_baseinfo=re.findall('class="ng-binding">(.*?)</p>',source)
# while company_baseinfo: #不知道为何无法实现换业 我的心在滴血。。。
print( company_baseinfo[0])
info['company_name']=company_baseinfo[0]
print( '公司名称:'+company_baseinfo[1])
info['Registered_Capital']=company_baseinfo[1].replace(' ','')
print( '注册资本:'+info['Registered_Capital'])
info['register_date']=company_baseinfo[2]
print( '注册时间:'+info['register_date'])
info['shareholder_info']=re.search('<meta name="description" content="(.*?)"',source,re.S).group(1)
print( '股东信息:'+info['shareholder_info'])
info['scope_of_business']=re.search('经营范围:</span>([\s\S]*)</p><!-- end ngIf: company.baseInfo.businessScope -->',source).group(1)
print( '经营范围:'+info['scope_of_business'])
info['register_place']=re.search('注册地址:</span>([\s\S]*)</p><!-- end ngIf: company.baseInfo.regLocation -->',source,re.S).group(1)
print( '注册地址:'+info['register_place'])
# info['conection_info']=re.search('<span class="contact_way_title">邮箱:</span>([\s\S]*)@qq.com',source,re.S).group(1) 如果抓取为空就会影响整个程序运行。。。
# print( '联系方式'+info['conection_info']
return info
except IndexError:
print(('No organization match for {}'))
def saveinfo(self,companyinfo):
f=open('jiangxiHr.txt','a')
for each in companyinfo:
f.writelines('公司名称:'+each['company_name']+'\n')
f.writelines('注册资本:'+each['Registered_Capital']+'\n')
f.writelines('注册时间:'+each['register_date']+'\n')
f.writelines('股东信息:'+each['shareholder_info']+'\n')
f.writelines('经营范围:'+each['scope_of_business']+'\n')
f.writelines('注册地址:'+each['register_place']+'\n')
f.close()
if __name__ == '__main__':
pool = ThreadPool(4)
classinfo = []
HRspider = spider()
for i in range(1,15):
url='http://www.tianyancha.com/search/%E6%B1%9F%E8%A5%BF%20%20%20%E4%BA%BA%E5%8A%9B%E8%B5%84%E6%BA%90/'+ str(i)
print( u'正在处理页面:' + url)
html=HRspider.get_source(url)
get_companylink=HRspider.get_companyurl(html)
for eachlink in get_companylink[1:19]:
companysource=HRspider.get_source(eachlink)
companyinfo=HRspider.get_companyinfo(companysource)
classinfo.append(companyinfo)
HRspider.saveinfo(classinfo)
pool.close()
pool.join() | [
"1938036263@qq.com"
] | 1938036263@qq.com |
6daee3b2a326fe785f8f51e527b35085f9128fff | 9f2577f0e5e0fc1a5e7f2d997ddac91be359f019 | /python-web/p39.py | 2a1909cb54791f7fcc26f17b28862f7f4c586cf2 | [] | no_license | diqiu11/python | f82cc02a2065b0f933b4d47b352e0f14072350d4 | 7af93ac8eed8667b3be26f4490d74f0aa74924ae | refs/heads/master | 2022-12-11T15:26:09.091309 | 2018-10-17T14:29:01 | 2018-10-17T14:29:01 | 149,900,411 | 0 | 1 | null | null | null | null | UTF-8 | Python | false | false | 170 | py | #!/usr/bin/env python
# coding=utf-8
import pickle
d = dict(name='bod',age=20)
print(pickle.dumps(d))
pickle.dumps(d)
f = open('test.txt','wb')
pickle.dump(d,f)
f.close
| [
"di15218143233@outlook.com"
] | di15218143233@outlook.com |
1575b08a2c652e7cdf3d3da4db1c9005fb2a2b5b | 3da6b8a0c049a403374e787149d9523012a1f0fc | /Coder_Old/几个好玩有趣的Python入门实例/简单统计/main.py | 36fcf153c6ae58736a502713a8e34905eff3b104 | [] | no_license | AndersonHJB/PyCharm_Coder | d65250d943e84b523f022f65ef74b13e7c5bc348 | 32f2866f68cc3a391795247d6aba69a7156e6196 | refs/heads/master | 2022-07-25T11:43:58.057376 | 2021-08-03T02:50:01 | 2021-08-03T02:50:01 | 348,922,058 | 3 | 3 | null | 2021-09-05T02:20:10 | 2021-03-18T02:57:16 | Python | UTF-8 | Python | false | false | 1,270 | py | # 输入一组数据,计算均值,方差,中位数,绝对相对误差。
# -*- coding: utf-8 -*-
# 输入数据
def getNum():
nums = []
iNumStr = input('please input a sequence of numbers (enter to exit): ')
while iNumStr != '':
nums.append(eval(iNumStr))
iNumStr = input('please input a sequence of numbers (enter to exit): ')
return nums
# 平均数
def average(numbers):
return sum(numbers) / len(numbers)
# 标准差
def dev(numbers, average):
sdev = 0.0
for num in numbers:
sdev += (num - average) ** 2
return pow(sdev / len(numbers), 0.5)
# 中位数
def median(numbers):
sorted(numbers)
size = len(numbers)
if size % 2 == 0:
return (numbers[size // 2 - 1] + numbers[size // 2]) / 2
else:
return numbers[size // 2]
# 绝对与相对误差
def rel_dev(numbers, average):
_max = max(abs(max(numbers) - average), abs(min(numbers) - average))
return _max, _max / average
def main():
nums = getNum()
if len(nums) == 0:
print('no data')
else:
ave = average(nums)
devs = rel_dev(nums, ave)
print('和:{:.4f},平均数:{:.4f},中位数:{:.4f},方差:{:.4f},绝对误差:{:4f},相对误差:{:.4f}' \
.format(sum(nums), ave, median(nums), dev(nums, ave), devs[0], devs[1]))
if __name__ == '__main__':
main()
| [
"1432803776@qq.com"
] | 1432803776@qq.com |
b5bf55019dfe3ec3859227a556b6f25df7dab10f | d89280c2cef8c17fc34a3a82a77f5bdb0e4c07a7 | /Veterinaria/wsgi.py | 3c14942e3f1266f018f2ab104d1a3dc38c0d3d3b | [] | no_license | Alejandro-Gutierrez/Vetesoftv2 | 821593233f70bf7b7d816157cd29ab41c645b58d | 84380ea51132f71d685c9b4b1f5f3a92a7a973b5 | refs/heads/master | 2022-06-20T06:36:44.633345 | 2019-09-19T20:57:55 | 2019-09-19T20:57:55 | 223,872,623 | 0 | 0 | null | 2019-11-25T05:59:32 | 2019-11-25T05:59:30 | null | UTF-8 | Python | false | false | 399 | py | """
WSGI config for Veterinaria project.
It exposes the WSGI callable as a module-level variable named ``application``.
For more information on this file, see
https://docs.djangoproject.com/en/2.1/howto/deployment/wsgi/
"""
import os
from django.core.wsgi import get_wsgi_application
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'Veterinaria.settings')
application = get_wsgi_application()
| [
"cristopherduarte50@gmail.com"
] | cristopherduarte50@gmail.com |
2bc90326c40d23b700ea9c755c27c0b0e2ed9867 | 138909a17b9f4b82ec91a209443864fbd18c1248 | /ExtraLongFactorial.py | f91e86d87985bb9d190fbd2252dde385422c8a42 | [] | no_license | surbhilakhani/Hackerrank | 70fc0a7bf85e73dbc6bd1f4695e148f7080a0c59 | f6cea99c5787c10ea5817bb9c4f3be8da1f6a73c | refs/heads/master | 2021-01-19T03:03:05.435417 | 2016-07-01T13:45:19 | 2016-07-01T13:45:19 | 62,326,553 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 82 | py | import sys
import math
n = int(raw_input().strip())
print math.factorial(n)
| [
"noreply@github.com"
] | noreply@github.com |
20a2640e2ad54b344e5be1bcbd8dfe4f8745ed6b | 55c250525bd7198ac905b1f2f86d16a44f73e03a | /Python/Games/Chess-py/gui/gui_functions.py | ee41fde1220a24a6a79a27e9b11f9b5729a73a9c | [] | no_license | NateWeiler/Resources | 213d18ba86f7cc9d845741b8571b9e2c2c6be916 | bd4a8a82a3e83a381c97d19e5df42cbababfc66c | refs/heads/master | 2023-09-03T17:50:31.937137 | 2023-08-28T23:50:57 | 2023-08-28T23:50:57 | 267,368,545 | 2 | 1 | null | 2022-09-08T15:20:18 | 2020-05-27T16:18:17 | null | UTF-8 | Python | false | false | 129 | py | version https://git-lfs.github.com/spec/v1
oid sha256:920dea71adf194f81da15c63d5ab5246c6637ed6329661630abdf4d56b12f7a6
size 9635
| [
"nateweiler84@gmail.com"
] | nateweiler84@gmail.com |
f98441cac34e408cb07f0ac6db2686cacde92265 | 35d979d4dd1110525fd4c31a78db147d59ec585d | /contact/admin.py | f08502ddc58c7319cf1090cc3a729abfaa7e1e05 | [] | no_license | Code-Institute-Submissions/gymnazium | 0eb32a1e61bde381e7c3716be2a99c9004c67d65 | 4161fafc4ffd6bf37e3a430c169defb19f02b04d | refs/heads/master | 2023-04-13T16:09:01.786918 | 2021-04-28T18:55:59 | 2021-04-28T18:55:59 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 307 | py | from django.contrib import admin
from.models import Contact
class ContactAdmin(admin.ModelAdmin):
list_display = ('name', 'email', 'phone', 'message', 'contact_date')
list_display_links = ('name',)
search_fields = ('name',)
list_per_page = 25
admin.site.register(Contact, ContactAdmin)
| [
"coreyhoward@live.co.uk"
] | coreyhoward@live.co.uk |
b001375d4646517af6ed057a18134e3829056b8a | ffbf665d491cafdfd88e10cea3628375ac7e7dbe | /mysite/env/bin/easy_install | 4651492f793dc04ad5c5479b0c61d850ad7c5a19 | [] | no_license | mnbqwe10/blackmores | fc885a6813ff6a3ba75da64e799612923e182669 | c3430727c9586a169e19179b430f75e4bd870e9b | refs/heads/master | 2021-04-06T00:15:27.314049 | 2018-04-11T14:21:06 | 2018-04-11T14:21:06 | 124,758,109 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 257 | #!/home/qqlivewell/mysite/env/bin/python3
# -*- coding: utf-8 -*-
import re
import sys
from setuptools.command.easy_install import main
if __name__ == '__main__':
sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
sys.exit(main())
| [
"qulong627@gmail.com"
] | qulong627@gmail.com | |
050d720d2e8a80dea21839c64f966f5092d892a2 | 1f2a5ca2f1631d30cd1a6b7916a13ec1299352d7 | /treasure/my_progress/apps.py | ae53c090de3372e80b757046c0b91d6a33592d6b | [] | no_license | danyyoo3/Personal-Website | 5449349e27e38865a10f06143c8c384c4be5a63b | b24ffd0c7d4fa311105fd158f459f98b3b8ef14d | refs/heads/main | 2023-03-03T13:33:07.408336 | 2021-02-15T04:03:47 | 2021-02-15T04:03:47 | 338,951,748 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 101 | py | from django.apps import AppConfig
class MyProgressConfig(AppConfig):
name = 'my_progress'
| [
"noreply@github.com"
] | noreply@github.com |
72b950570c7fa0784332ef7fa8b80fda93f2945a | 104d097ee850d8643b97d6bc17722e322400c358 | /desafio6.py | 16e05c478f7c8d95ac97577e322d925cc1717659 | [] | no_license | rodolfocr92/desafio | 8ba1b031b1f8b3e266af62a7d0b2678ffa0dda14 | 65cc9570797e33b885dc0d21252ecfd674774428 | refs/heads/master | 2022-11-17T22:24:48.713745 | 2020-07-12T19:01:08 | 2020-07-12T19:01:08 | 277,664,948 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 396 | py |
io = float(input("Insira sua idade atual: "))
ia = float(input("Insira a idade que deseja sacar o investimento: "))
vf = float(input("Insira o montante final do investimento: "))
i = float(input("Insira a taxa de juros mensais em número decimal: "))
meses = (ia - io)*12
dep = (vf*i)/((1+i)**meses - 1)
depr = round(dep, 2)
print("Valor do depósito mensal: R$" + str(depr))
| [
"noreply@github.com"
] | noreply@github.com |
05c708c3a242b5a6ac1b32a631a8455d98419b06 | 36a8ac888405e310c551713de49b6ceee9416257 | /automate_boring_stuff/isPhoneNumber.py | 9458cb2429c012943bd7d9b8c70fb2ed45f05c60 | [] | no_license | gonkun/programming_practices | 8a0615e96714565a347973aeba813d4c79814c46 | 41757e9e04de7c6c6e1e1e408072800d1f5eb312 | refs/heads/master | 2023-02-23T07:39:52.024638 | 2022-04-08T10:39:41 | 2022-04-08T10:39:41 | 149,608,231 | 0 | 0 | null | 2023-02-16T00:59:09 | 2018-09-20T12:47:30 | Python | UTF-8 | Python | false | false | 661 | py | def isPhoneNumber(text):
if len(text) != 12:
return False
for i in range(0, 3):
if not text[i].isdecimal():
return False
if text[3] != "-":
return False
for i in range(4, 7):
if not text[i].isdecimal():
return False
if text[7] != "-":
return False
for i in range(8, 12):
if not text[i].isdecimal():
return False
return True
message = "Call me at 415-555-1011 tomorrow. 415-555-9999 is my office."
for i in range(len(message)):
chunk = message[i: i + 12]
if isPhoneNumber(chunk):
print("Phone number found: " + chunk)
print("Done")
| [
"gonzalo.sanchter@gmail.com"
] | gonzalo.sanchter@gmail.com |
6bbe246fd9bd6d0eb23ccd5e2f43f5280487874c | d29c2dea4afbb21de0b1e508e501ee6711805451 | /__main__.py | e084aa8b11fab88e422d61a1e430451cb2602f83 | [
"MIT"
] | permissive | cdeitrick/workflows | ef69003cbd6030bc828815b7c898128327da129a | 8edd2a08078144a2445af3903eb13b71abb96538 | refs/heads/master | 2020-03-18T07:04:20.554986 | 2019-12-18T21:16:39 | 2019-12-18T21:16:39 | 134,430,686 | 0 | 0 | MIT | 2019-07-11T03:29:48 | 2018-05-22T14:50:28 | Python | UTF-8 | Python | false | false | 333 | py | from pipelines import main
import argparse
def create_parser()->argparse.Namespace:
parser = argparse.ArgumentParser()
parser.add_argument(
"which",
help = "assembly or variants.",
type = str,
choices = ['assembly', 'variants']
)
args = parser.parse_args()
return args
if __name__ == "__main__":
main.main_shelly() | [
"cld100@pitt.edu"
] | cld100@pitt.edu |
7045ae2111e975c1900c7f15ec0532dbbf283c3d | 9a076ee891aa04dd1522662838dda63ad554e835 | /manage.py | 6c345e52c5342bfb1e480ee19abe93787dd7e988 | [
"MIT"
] | permissive | Albert-Byrone/Pitches | fc018b3f46ea325456212154f27426c7d18ef435 | d9ae032ff0a00b135d03404477e07a8405247b5e | refs/heads/master | 2022-10-16T14:46:38.758963 | 2019-10-22T11:53:35 | 2019-10-22T11:53:35 | 216,051,524 | 0 | 1 | MIT | 2022-09-16T18:11:23 | 2019-10-18T15:12:54 | Python | UTF-8 | Python | false | false | 619 | py | from flask_script import Manager, Server
from flask_migrate import Migrate, MigrateCommand
from app import create_app,db
from app.models import User
app = create_app('production')
manager = Manager(app)
migrate = Migrate(app,db)
manager.add_command('db',MigrateCommand)
manager.add_command('server',Server(use_debugger=True))
@manager.shell
def make_shell_context():
return dict(app = app,db = db,User = User)
@manager.command
def test():
import unittest
tests = unittest.TestLoader().discover('tests')
unittest.TextTestRunner(verbosity=2).run(tests)
if __name__ == "__main__":
manager.run()
| [
"albertbyrone1677@gmail.com"
] | albertbyrone1677@gmail.com |
c74a079a4da583c78d6b6231f9aed74ecc76caab | ee543767f2af3113645adc2db6ffa50029a1b58e | /tests/dataset/convert_dataset.py | 6c82d5adcab5c9ef2c9c821e103eab1662975cc9 | [
"MIT"
] | permissive | parkertomf/WrightTools | 97bb466d2a83512120b43d5aa87a034a92f94db9 | 86c6f71fc4d157f4d112fb94e5b8055246912dc8 | refs/heads/master | 2020-03-18T22:20:38.587304 | 2018-06-11T23:11:20 | 2018-06-11T23:11:20 | 135,339,820 | 0 | 0 | MIT | 2018-05-29T18:52:58 | 2018-05-29T18:52:57 | null | UTF-8 | Python | false | false | 1,245 | py | """Test dataset unit conversion."""
# --- import --------------------------------------------------------------------------------------
import numpy as np
import WrightTools as wt
from WrightTools import datasets
# --- define --------------------------------------------------------------------------------------
def test_exception():
p = datasets.PyCMDS.w1_000
data = wt.data.from_PyCMDS(p)
try:
data['w1'].convert('fs')
except wt.exceptions.UnitsError:
assert True
else:
assert False
assert data['w1'].units == 'nm'
data.close()
def test_w1_wa():
p = datasets.PyCMDS.w1_wa_000
data = wt.data.from_PyCMDS(p)
assert data['wa'].units == 'nm'
data['wa'].convert('eV')
assert np.isclose(data['wa'].max(), 1.5802564757220569)
assert np.isclose(data['wa'].min(), 0.6726385958618104)
data.close()
def test_wigner():
p = datasets.COLORS.v2p2_WL_wigner
data = wt.data.from_COLORS(p)
data['d1'].convert('ns')
assert data['d1'].units == 'ns'
data.close()
# --- run -----------------------------------------------------------------------------------------
if __name__ == '__main__':
test_exception()
test_w1_wa()
test_wigner()
| [
"ksunden@users.noreply.github.com"
] | ksunden@users.noreply.github.com |
7f43ad6eb669411e466e915a50d1333fea97f289 | dd2f95416ec8107680fe72dc997eaff55fb79698 | /PixelDBTools/test/SiPixelLorentzAngleDBLoader_cfg.py | 61b6392c97fa0aedacc997546de0db95a1cb6505 | [] | no_license | odysei/DPGAnalysis-SiPixelTools | f433e75a642aea9dbc5574e86a8359826e83e556 | ac4fdc9fa3ad8a865acb64dc3bbefe66b1bdd45a | refs/heads/master | 2021-01-18T05:55:15.102609 | 2015-09-15T13:10:41 | 2015-09-15T13:10:41 | 42,519,902 | 0 | 0 | null | 2015-09-15T13:07:20 | 2015-09-15T13:07:19 | null | UTF-8 | Python | false | false | 9,799 | py | #
import FWCore.ParameterSet.Config as cms
process = cms.Process("SiPixelLorentzAngleLoader")
process.load("Configuration.Geometry.GeometryRecoDB_cff")
process.load("Configuration.StandardSequences.MagneticField_cff")
#process.load("Geometry.CMSCommonData.cmsIdealGeometryXML_cfi")
process.load("CalibTracker.Configuration.TrackerAlignment.TrackerAlignment_Fake_cff")
#process.load("Geometry.TrackerGeometryBuilder.trackerGeometry_cfi")
#process.load("Geometry.TrackerNumberingBuilder.trackerNumberingGeometry_cfi")
process.load("CondTools.SiPixel.SiPixelGainCalibrationService_cfi")
process.load("CondCore.DBCommon.CondDBCommon_cfi")
process.load("Configuration.StandardSequences.FrontierConditions_GlobalTag_cff")
#process.load("Configuration.StandardSequences.FrontierConditions_GlobalTag_condDBv2_cff")
from Configuration.AlCa.GlobalTag_condDBv2 import GlobalTag
#from Configuration.AlCa.GlobalTag import GlobalTag
#process.GlobalTag = GlobalTag(process.GlobalTag, 'auto:run2_data', '')
#process.GlobalTag = GlobalTag(process.GlobalTag, 'auto:run1_data', '')
process.GlobalTag = GlobalTag(process.GlobalTag, 'auto:run2_mc', '')
#process.GlobalTag = GlobalTag(process.GlobalTag, 'auto:run2_design', '')
process.load("FWCore.MessageService.MessageLogger_cfi")
process.MessageLogger.destinations = cms.untracked.vstring("cout")
process.MessageLogger.cout = cms.untracked.PSet(threshold = cms.untracked.string("ERROR"))
process.source = cms.Source("EmptyIOVSource",
firstValue = cms.uint64(1),
lastValue = cms.uint64(1),
timetype = cms.string('runnumber'),
interval = cms.uint64(1)
)
process.maxEvents = cms.untracked.PSet(
input = cms.untracked.int32(1)
)
# has to be deleted if it exist
file = "la.db"
sqlfile = "sqlite_file:" + file
print '\n-> Uploading into file %s, i.e. %s\n' % (file, sqlfile)
##### DATABASE CONNNECTION AND INPUT TAGS ######
process.PoolDBOutputService = cms.Service("PoolDBOutputService",
BlobStreamerName = cms.untracked.string('TBufferBlobStreamingService'),
DBParameters = cms.PSet(
authenticationPath = cms.untracked.string('.'),
connectionRetrialPeriod = cms.untracked.int32(10),
idleConnectionCleanupPeriod = cms.untracked.int32(10),
messageLevel = cms.untracked.int32(1),
enablePoolAutomaticCleanUp = cms.untracked.bool(False),
enableConnectionSharing = cms.untracked.bool(True),
connectionRetrialTimeOut = cms.untracked.int32(60),
connectionTimeOut = cms.untracked.int32(0),
enableReadOnlySessionOnUpdateConnection = cms.untracked.bool(False)
),
timetype = cms.untracked.string('runnumber'),
connect = cms.string(sqlfile),
toPut = cms.VPSet(
cms.PSet(
record = cms.string('SiPixelLorentzAngleRcd'),
tag = cms.string('SiPixelLorentzAngle_test')
# tag = cms.string("SiPixelLorentzAngle_fromAlignment_v01_mc")
# tag = cms.string("SiPixelLorentzAngle_fromAlignment_v01")
# tag = cms.string("SiPixelLorentzAngle_forWidth_v01_mc")
# tag = cms.string("SiPixelLorentzAngle_forWidth_v01")
),
cms.PSet(
record = cms.string('SiPixelLorentzAngleSimRcd'),
tag = cms.string('SiPixelLorentzAngleSim_test')
),
)
)
###### LORENTZ ANGLE OBJECT ######
process.SiPixelLorentzAngle = cms.EDAnalyzer("SiPixelLorentzAngleDBLoader",
# common input for all rings
bPixLorentzAnglePerTesla = cms.double(0.10),
fPixLorentzAnglePerTesla = cms.double(0.06),
# bPixLorentzAnglePerTesla = cms.double(0.05),
# fPixLorentzAnglePerTesla = cms.double(0.03),
# enter -9999 if individual input for rings
# bPixLorentzAnglePerTesla = cms.double(-9999.),
# fPixLorentzAnglePerTesla = cms.double(-9999.),
#in case of PSet
BPixParameters = cms.untracked.VPSet(
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(1),
angle = cms.double(0.0948)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(2),
angle = cms.double(0.0948)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(3),
angle = cms.double(0.0948)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(4),
angle = cms.double(0.0948)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(5),
angle = cms.double(0.0964)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(6),
angle = cms.double(0.0964)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(7),
angle = cms.double(0.0964)
),
cms.PSet(
layer = cms.uint32(1),
module = cms.uint32(8),
angle = cms.double(0.0964)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(1),
angle = cms.double(0.0916)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(2),
angle = cms.double(0.0916)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(3),
angle = cms.double(0.0916)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(4),
angle = cms.double(0.0916)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(5),
angle = cms.double(0.0931)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(6),
angle = cms.double(0.0931)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(7),
angle = cms.double(0.0931)
),
cms.PSet(
layer = cms.uint32(2),
module = cms.uint32(8),
angle = cms.double(0.0931)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(1),
angle = cms.double(0.0920)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(2),
angle = cms.double(0.0920)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(3),
angle = cms.double(0.0920)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(4),
angle = cms.double(0.0920)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(5),
angle = cms.double(0.0935)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(6),
angle = cms.double(0.0935)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(7),
angle = cms.double(0.0935)
),
cms.PSet(
layer = cms.uint32(3),
module = cms.uint32(8),
angle = cms.double(0.0935)
),
),
FPixParameters = cms.untracked.VPSet(
cms.PSet(
side = cms.uint32(1),
disk = cms.uint32(1),
HVgroup = cms.uint32(1),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(1),
disk = cms.uint32(2),
HVgroup = cms.uint32(1),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(2),
disk = cms.uint32(1),
HVgroup = cms.uint32(1),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(2),
disk = cms.uint32(2),
HVgroup = cms.uint32(1),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(1),
disk = cms.uint32(1),
HVgroup = cms.uint32(2),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(1),
disk = cms.uint32(2),
HVgroup = cms.uint32(2),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(2),
disk = cms.uint32(1),
HVgroup = cms.uint32(2),
angle = cms.double(0.081)
),
cms.PSet(
side = cms.uint32(2),
disk = cms.uint32(2),
HVgroup = cms.uint32(2),
angle = cms.double(0.081)
),
),
#in case lorentz angle values for bpix should be read from file -> not implemented yet
useFile = cms.bool(False),
record = cms.untracked.string('SiPixelLorentzAngleRcd'),
fileName = cms.string('lorentzFit.txt')
)
process.SiPixelLorentzAngleSim = cms.EDAnalyzer("SiPixelLorentzAngleDBLoader",
# magneticField = cms.double(3.8),
bPixLorentzAnglePerTesla = cms.double(0.10),
fPixLorentzAnglePerTesla = cms.double(0.06),
#in case lorentz angle values for bpix should be read from file -> not implemented yet
useFile = cms.bool(False),
record = cms.untracked.string('SiPixelLorentzAngleSimRcd'),
fileName = cms.string('lorentzFit.txt'),
#in case of PSet
BPixParameters = cms.untracked.VPSet(
cms.PSet(
layer = cms.uint32(0),
module = cms.uint32(0),
angle = cms.double(0.0)
),
),
FPixParameters = cms.untracked.VPSet(
cms.PSet(
side = cms.uint32(0),
disk = cms.uint32(0),
HVgroup = cms.uint32(0),
angle = cms.double(0.0)
),
),
)
process.p = cms.Path(
# process.SiPixelLorentzAngleSim
process.SiPixelLorentzAngle
)
| [
"danek.kotlinski@psi.ch"
] | danek.kotlinski@psi.ch |
977e71a983cd038695cbd3ea08c4705d1f871367 | 9671c2b41ff71727d72fd7f43006a2a968561200 | /local/neuralNetworks/classifiers/lstm.py | f3108e7903c37548e9740b50908d34460233f2ab | [] | no_license | didadida-r/tf-cnn | 4adb6ceefad6d828e8e847bfe3680bc8095e17b1 | 8cc39bdace55630a999607c19091dce2342281a9 | refs/heads/master | 2023-01-10T22:51:51.771840 | 2017-06-03T04:11:55 | 2017-06-03T04:11:55 | null | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 7,147 | py | '''@file lstm.py
The LSTM neural network classifier'''
import seq_convertors
import tensorflow as tf
from classifier import Classifier
from layer import FFLayer
from activation import TfActivation
import inspect
class LSTM(Classifier):
'''This class is a graph for lstm neural nets.'''
def __init__(self, output_dim, num_layers, num_units, activation,
layerwise_init=True):
'''
DNN constructor
Args:
output_dim: the DNN output dimension
num_layers: number of hidden layers
num_units: number of hidden units
activation: the activation function
layerwise_init: if True the layers will be added one by one,
otherwise all layers will be added to the network in the
beginning
'''
#super constructor
super(LSTM, self).__init__(output_dim)
#save all the DNN properties
self.num_layers = num_layers
print(self.num_layers)
self.num_units = num_units
print(self.num_units)
self.activation = activation
self.layerwise_init = layerwise_init
self.layerwise_init = None
def __call__(self, inputs, seq_length, is_training=False, reuse=False,
scope=None):
'''
Add the LSTM variables and operations to the graph
Args:
inputs: the inputs to the neural network, this is a list containing
a [batch_size, input_dim] tensor for each time step
seq_length: The sequence lengths of the input utterances, if None
the maximal sequence length will be taken
is_training: whether or not the network is in training mode
reuse: wheter or not the variables in the network should be reused
scope: the name scope
Returns:
A triple containing:
- output logits
- the output logits sequence lengths as a vector
- a saver object
- a dictionary of control operations:
-add: add a layer to the network
-init: initialise the final layer
'''
with tf.variable_scope(scope or type(self).__name__, reuse=reuse):
weights = {'out':
tf.get_variable('weights_out', [self.num_units, self.output_dim], initializer=tf.contrib.layers.xavier_initializer())
}
biases = {'out':
tf.get_variable('biases_out', [self.output_dim], initializer=tf.constant_initializer(0))
}
#convert the sequential data to non sequential data
nonseq_inputs = seq_convertors.seq2nonseq(inputs, seq_length)
input_dim = nonseq_inputs.shape[1]
nonseq_inputs = tf.reshape(nonseq_inputs,[-1,11,40])
n_steps = 11
nonseq_inputs = tf.transpose(nonseq_inputs, [1, 0, 2])
keep_prob = 1
# define the lstm cell
# use the dropout in training mode
if is_training and keep_prob < 1:
lstm_cell = tf.contrib.rnn.LayerNormBasicLSTMCell(self.num_units, forget_bias=0.0,
input_size=None, activation=tf.nn.relu, layer_norm=False, norm_gain=1.0,
norm_shift=0.0, dropout_keep_prob=keep_prob, dropout_prob_seed=None)
lstm_cell = tf.contrib.rnn.LayerNormBasicLSTMCell(self.num_units, forget_bias=0.0,
input_size=None, activation=tf.nn.relu, layer_norm=False, norm_gain=1.0,
norm_shift=0.0, dropout_keep_prob=1, dropout_prob_seed=None)
# stack the lstm to form multi-layers
cell = tf.contrib.rnn.MultiRNNCell(
[lstm_cell]*self.num_layers, state_is_tuple=True)
# print(int(nonseq_inputs.shape[0]))
# self._initial_state = cell.zero_state(int(nonseq_inputs.shape[0]), tf.float32)
# apply the dropout for the inputs to the first hidden layer
if is_training and keep_prob < 1:
nonseq_inputs = tf.nn.dropout(nonseq_inputs, keep_prob)
final_nonseq_inputs = tf.unstack(nonseq_inputs, num=n_steps, axis=0)
# Get lstm cell output initial_state=self._initial_state,
outputs, states = tf.contrib.rnn.static_rnn(cell, final_nonseq_inputs, dtype=tf.float32)
outputs = outputs[-1]
# Linear activation, using rnn inner loop last output
logits = tf.matmul(outputs, weights['out']) + biases['out']
# # if self.layerwise_init:
# # #variable that determines how many layers are initialised
# # #in the neural net
# # initialisedlayers = tf.get_variable(
# # 'initialisedlayers', [],
# # initializer=tf.constant_initializer(0),
# # trainable=False,
# # dtype=tf.int32)
# # #operation to increment the number of layers
# # add_layer_op = initialisedlayers.assign(initialisedlayers+1).op
# # #compute the logits by selecting the activations at the layer
# # #that has last been added to the network, this is used for layer
# # #by layer initialisation
# # logits = tf.case(
# # [(tf.equal(initialisedlayers, tf.constant(l)),
# # Callable(activations[l]))
# # for l in range(len(activations))],
# # default=Callable(activations[-1]),
# # exclusive=True, name='layerSelector')
# # logits.set_shape([None, self.num_units])
if self.layerwise_init:
#operation to initialise the final layer
init_last_layer_op = tf.initialize_variables(
tf.get_collection(
tf.GraphKeys.VARIABLES,
scope=(tf.get_variable_scope().name + '/layer'
+ str(self.num_layers))))
control_ops = {'add':add_layer_op, 'init':init_last_layer_op}
else:
control_ops = None
#convert the logits to sequence logits to match expected output
seq_logits = seq_convertors.nonseq2seq(logits, seq_length,
len(inputs))
#create a saver
saver = tf.train.Saver()
return seq_logits, seq_length, saver, control_ops
class Callable(object):
'''A class for an object that is callable'''
def __init__(self, value):
'''
Callable constructor
Args:
tensor: a tensor
'''
self.value = value
def __call__(self):
'''
get the object
Returns:
the object
'''
return self.value | [
"605467297@qq.com"
] | 605467297@qq.com |
346ab1f2ac22d28a59f677348f56c7887974c730 | e22a4330d906f51932bf55eeeba705f447b74699 | /python/HorizonPlot.py | e8240f75787975dcc9c2d56e122a9f5cafffe902 | [
"Apache-2.0"
] | permissive | titodalcanton/flare | 5dc39c18152205b1fe2f1a0ed4f7e0607adf6fd4 | 4ffb02977d19786ab8c1a767cc495a799d9575ae | refs/heads/master | 2020-03-29T06:21:10.932174 | 2018-07-23T15:51:24 | 2018-07-23T15:51:24 | 149,621,008 | 0 | 0 | Apache-2.0 | 2018-09-20T14:19:13 | 2018-09-20T14:19:13 | null | UTF-8 | Python | false | false | 2,020 | py | import flare
import numpy as np
import argparse
eps=0.3
parser = argparse.ArgumentParser(description="generate a discover queue script for a flare run");
parser.add_argument('label',help="The basename/directory for the run")
args=parser.parse_args()
flare.flare_dir="../flare"
#file="FisherStudy-most-HH.dat"
#file="FisherStudy.dat"
label=args.label+"/"
#label="test11.15"
#label="L3LISA-v1-sens-but-5Gm-test-wide"
#label="L3LISARef"
#label="LISA2017camp_10yr/"
#label="2arm-LISA/"
#label="tRef-redef-LISA2017/"
#label="LISA2017-Nov-flaretest/"
#label="slow-orbit-LISA/"
#label="fast-orbit-LISA/"
#label="big-orbit-LISA/"
#label="slower-orbit-LISA/"
#label="tiny-orbit-LISA/"
file=label+"FisherStudy.dat"
#flare.FisherPlot(label,9,[1.1,2.0,4.0,10.0],[10,100,1000],[0.1,0.3],file)
#flare.FisherPlot(label,0,[1.1,2,4,10],[10,100,1000],[0.1,.3],file)
#flare.FisherPlot(label,0,[2],[10],[.3],file)
flare.HorizonPlot(label,0,[2],10,eps,file,[0.001,0.003,0.01,0.03,0.10,0.30],scaled=True)
flare.HorizonPlot(label,0,[2],10,eps,file,[0.001,0.01,0.10],scaled=True,show_range=True)
flare.HorizonPlot(label,1,[2],10,eps,file,[0.001,0.003,0.01,0.03,0.10,0.30],scaled=True)
flare.HorizonPlot(label,1,[2],10,eps,file,[0.001,0.01,0.10],scaled=True,show_range=True)
#flare.HorizonPlot(label,2,[2],10,[.3],file,[0.10,0.30,1.0,3.0,10.0,30.0,100.0],scaled=True)
flare.HorizonPlot(label,3,[2],10,eps,file,[0.003,0.01,0.03,0.10,0.30],scaled=True)
flare.HorizonPlot(label,3,[2],10,eps,file,[0.01,0.10,0.5],scaled=True,show_range=True)
#flare.HorizonPlot(label,9,[2],10,[.3],file,[8.4e-7,8.4e-6,8.4e-5,3.0e-4,3.0e-3,3.0e-2],scaled=True)
#flare.HorizonPlot(label,9,[2],10,[.3],file,[100,900,3600,28800,360000],scaled=True)
#flare.HorizonPlot(label,9,[2],10,[.3],file,[100,3600,90000],scaled=True,show_range=True)
flare.HorizonPlot(label,9,[2],10,eps,file,[0.0278,0.25,1.0,9.0,100.0],scaled=True)
flare.HorizonPlot(label,9,[2],10,eps,file,[0.0278,1.0,25.0],scaled=True,show_range=True)
#flare.FisherPlot(label,0,[2],10,[.3],file,scaled=False)
| [
"john.g.baker@nasa.gov"
] | john.g.baker@nasa.gov |
25b944adffbb62adf37669bd50cacc9ad4b63b47 | c6c61add7e33535e16bc2d6c53cc482ecf11262e | /manage.py | b7a8e3f0aadadd79d752e4aa271f02f5fabce82c | [] | no_license | dfreidin/DjangoGameNight | 300e95eb849f3588a72e415837fffa5091ca8ba3 | 2880c09f3d45a74089823d1abde1401ffa775781 | refs/heads/master | 2020-03-07T03:09:36.094097 | 2018-04-14T21:28:44 | 2018-04-14T21:28:44 | 127,227,841 | 0 | 0 | null | null | null | null | UTF-8 | Python | false | false | 807 | py | #!/usr/bin/env python
import os
import sys
if __name__ == "__main__":
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "GameNight.settings")
try:
from django.core.management import execute_from_command_line
except ImportError:
# The above import may fail for some other reason. Ensure that the
# issue is really that Django is missing to avoid masking other
# exceptions on Python 2.
try:
import django
except ImportError:
raise ImportError(
"Couldn't import Django. Are you sure it's installed and "
"available on your PYTHONPATH environment variable? Did you "
"forget to activate a virtual environment?"
)
raise
execute_from_command_line(sys.argv)
| [
"36039896+dfreidin@users.noreply.github.com"
] | 36039896+dfreidin@users.noreply.github.com |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.