blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
75ad2856c5ff6b4ca39329ba99881252b8c694e8
6140bcfbdfad6fef444acfd8d40a776f7f467598
/sbmOpenMM/datasets/__init__.py
1820e9b8e9b585d1513c82d6822d7f0a0098220f
[ "MIT" ]
permissive
CompBiochBiophLab/sbm-openmm
bfa799f305d32b3b9de147d99bf8f2354867ee06
113b04eb2c78664a48600a05ff6778898fbb2fd9
refs/heads/master
2022-08-10T00:49:30.874501
2022-07-24T11:12:31
2022-07-24T11:12:31
219,484,731
18
7
MIT
2019-11-26T12:59:44
2019-11-04T11:23:38
Jupyter Notebook
UTF-8
Python
false
false
28
py
from . import foxp1_folding
[ "martin.floor@uvic.cat" ]
martin.floor@uvic.cat
2b342274bc429eacca9e41da3bbf739877dd0db6
a1ff3f4a3909b672a31338aa7a1e25d7953111fb
/twitter_app/ttc_debug.py
a9c691ffbfd6b244b1ba4828a633dfd53ce6f660
[]
no_license
wayne927/RPi-stuff
0301831576bbc0d3e9a480d57d04279452dd2d84
cae95547e96e0b1e80f2dc3b6de9b7dfeef777b9
refs/heads/master
2020-05-21T04:36:02.799577
2018-12-11T11:00:02
2018-12-11T11:00:02
58,220,398
0
0
null
null
null
null
UTF-8
Python
false
false
2,788
py
#!/usr/bin/python import twitter import datetime from datetime import timedelta import calendar import os import sys from readkeys import * from send_gmail import * def getLocalTime(date_in) : utc = datetime.datetime.strptime(date_in, '%a %b %d %H:%M:%S +0000 %Y') # TODO: deal with daylight saving time... offset = -5 local_time = utc + timedelta(hours=offset) return local_time def shouldPrintTweet(status) : # Don't care if it's weekend (note: weekday() is 0 for Monday) local_time = getLocalTime(status.created_at) if local_time.weekday() >= 5 : return False # convert to lower cases, strip all white spaces tweet = ''.join(status.text.lower().split()) include_any = ['line1', 'line2', 'line4'] exclude_all = ['elevator'] inc = False for w in include_any : if w in tweet : inc = True break # Doesn't have any include keywords. Don't need to filter excludes if inc == False : return False for w in exclude_all : if w in tweet : return False return True def printTweet(status, output) : local_time = getLocalTime(status.created_at) out = "" #out = out + str(status.id) + '\n' out = out + calendar.day_name[local_time.weekday()] + " " out = out + str(local_time) + '\n' + status.text + '\n\n' return output + out keys = readKeys('binkeys.apikey') api = twitter.Api(consumer_key=keys[0], consumer_secret=keys[1], access_token_key=keys[2], access_token_secret=keys[3]) # read the most recent status (MRS) id that we got last time # try : # fileMRS = open('mrs.txt', 'r') # MRS_id = int(fileMRS.readline()) # fileMRS.close() # print('Most recent status ID read from file = ' + str(MRS_id)) # except : # # File not found? Bad id? Meh # MRS_id = 0 MRS_id = 1014528457500385280 if MRS_id == 0 : print('MRS ID invalid. Just read the last 100.') statuses = api.GetUserTimeline(screen_name='TTCnotices', count=100) MRS_id = statuses[0].id else : statuses = api.GetUserTimeline(screen_name='TTCnotices', since_id=MRS_id, count=1000) print('Number of statuses since last MRS = ' + str(len(statuses))) if len(statuses) == 0 : sys.exit() else : MRS_id = statuses[0].id # fileMRS = open('mrs.txt', 'w') # fileMRS.write(str(MRS_id)) # fileMRS.close() output = "" for s in statuses : tweet = ''.join(s.text.lower().split()) if shouldPrintTweet(s) : output = printTweet(s, output) if not output : sys.exit() #timenow = datetime.datetime.now() #email_subject = 'TTC Update: ' + timenow.strftime('%a %b %d %H:%M:%S') #send_gmail(email_subject, output) print(output)
[ "wayne.ngan@gmail.com" ]
wayne.ngan@gmail.com
43e3f69a4d43e8fd97a6995fa95b1197d002dc0e
0315255c749b12216a7c8ac26378d8921466284a
/tests/integration/client/standard.py
969611b4d0a0800f10b1c10258875138538f5b08
[ "Apache-2.0" ]
permissive
jhutchins/salt
a32de1362c6787ec96df7ce57bf9b98f20eaf30a
22ec0cee6a8a842ec426b7a3e634723ea7ce7256
refs/heads/master
2021-01-21T00:05:05.782149
2012-04-06T22:03:19
2012-04-06T22:03:19
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,254
py
# Import python libs import subprocess # Import salt libs import integration class StdTest(integration.ModuleCase): ''' Test standard client calls ''' def test_cli(self): ''' Test cli function ''' cmd_iter = self.client.cmd_cli( 'minion', 'test.ping', ) for ret in cmd_iter: self.assertTrue(ret['minion']) def test_iter(self): ''' test cmd_iter ''' cmd_iter = self.client.cmd_iter( 'minion', 'test.ping', ) for ret in cmd_iter: self.assertTrue(ret['minion']) def test_iter_no_block(self): ''' test cmd_iter_no_block ''' cmd_iter = self.client.cmd_iter_no_block( 'minion', 'test.ping', ) for ret in cmd_iter: if ret is None: continue self.assertTrue(ret['minion']) def test_full_returns(self): ''' test cmd_iter ''' ret = self.client.cmd_full_return( 'minion', 'test.ping', ) self.assertTrue(ret['minion'])
[ "thatch45@gmail.com" ]
thatch45@gmail.com
d96d507a3a852249066e493e07e09227d669afb0
cd8ea41cd625440bfacbee84e7a7632e1243c407
/code/recipe/tests/test_ingredients_api.py
e22fc54e3b2e1e1bab6a44b16230688e9876e504
[]
no_license
DamianCzajkowski/recipe-app
826c332f5b6699cb3f63c5005dda953928467e20
a593acbfedc6b0142f3e51289cc4a4dbd19bf793
refs/heads/main
2023-07-18T06:16:28.872407
2021-08-15T19:12:49
2021-08-15T19:12:49
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,346
py
from django.contrib.auth import get_user_model from django.urls import reverse from django.test import TestCase from rest_framework import status from rest_framework.test import APIClient from core.models import Ingredient, Recipe from recipe.serializers import IngredientSerializer INGREDIENTS_URL = reverse('recipe:ingredient-list') class PublicIngredientsApiTests(TestCase): """Test the publicly available ingredients API""" def setUp(self): self.client = APIClient() def test_login_required(self): """Test that login is required to access the endpoint""" response = self.client.post(INGREDIENTS_URL) self.assertEqual(response.status_code, status.HTTP_401_UNAUTHORIZED) class PrivateIngredientsApiTests(TestCase): """Test the private ingredients API""" def setUp(self): self.client = APIClient() self.user = get_user_model().objects.create_user('test@test.com', 'passwd123') self.client.force_authenticate(self.user) def test_retrieve_ingredients_list(self): """Test retrieving a list of ingredients""" Ingredient.objects.create(user=self.user, name="Kale") Ingredient.objects.create(user=self.user, name="Salt") response = self.client.get(INGREDIENTS_URL) ingredients = Ingredient.objects.all().order_by("-name") serializer = IngredientSerializer(ingredients, many=True) self.assertEqual(response.status_code, status.HTTP_200_OK) self.assertEqual(response.data, serializer.data) def test_ingredients_limited_to_user(self): """Test that ingredients for the authenticated user are returned""" user2 = get_user_model().objects.create_user('test2@test.com', 'passwd123') Ingredient.objects.create(user=user2, name='Vinegar') ingredient = Ingredient.objects.create(user=self.user, name='Tumeric') response = self.client.get(INGREDIENTS_URL) self.assertEqual(response.status_code, status.HTTP_200_OK) self.assertEqual(len(response.data), 1) self.assertEqual(response.data[0]['name'], ingredient.name) def test_create_ingredient_successful(self): """Test create a new ingredient""" payload = {"name": "Cabbage"} self.client.post(INGREDIENTS_URL, payload) exists = Ingredient.objects.filter( user=self.user, name=payload["name"]).exists() self.assertTrue(exists) def test_create_ingredient_invalid(self): """Test creating ingredients invalid""" payload = {"name": ""} response = self.client.post(INGREDIENTS_URL, payload) self.assertEqual(response.status_code, status.HTTP_400_BAD_REQUEST) def test_retrieve_ingredients_assigned_to_recipes(self): """Test filtering ingredients by those assigned to recipes""" ingredient1 = Ingredient.objects.create(user=self.user, name='Apples') ingredient2 = Ingredient.objects.create(user=self.user, name='Turkey') recipe = Recipe.objects.create( title='Apple crumble', time_minutes=5, price=10.00, user=self.user ) recipe.ingredients.add(ingredient1) response = self.client.get(INGREDIENTS_URL, {'assigned_only': 1}) serializer1 = IngredientSerializer(ingredient1) serializer2 = IngredientSerializer(ingredient2) self.assertIn(serializer1.data, response.data) self.assertNotIn(serializer2.data, response.data) def test_retrieve_ingredients_assigned_unique(self): """Test filtering ingredients by assigned returns unique items""" ingredient = Ingredient.objects.create(user=self.user, name='Eggs') Ingredient.objects.create(user=self.user, name='Cheese') recipe1 = Recipe.objects.create( title='Eggs benedict', time_minutes=30, price=12.00, user=self.user ) recipe1.ingredients.add(ingredient) recipe2 = Recipe.objects.create( title='Coriander eggs on toast', time_minutes=20, price=5.00, user=self.user ) recipe2.ingredients.add(ingredient) response = self.client.get(INGREDIENTS_URL, {'assigned_only': 1}) self.assertEqual(len(response.data), 1)
[ "damianczajkowski@bluerider.software" ]
damianczajkowski@bluerider.software
d341c94a528f020251ed8186095bde571f3fc3e8
e98d4ac448b838712616f8fe408a4073b799a478
/currency_converter_api.py
e1a8943aeb23ec5613a3b4583276540f55642fdc
[]
no_license
KachusovStepan/CurrencyConverterApi
2a15bbe0e5376480b4c752d4585611647d9f0f8d
485460ee9a0c04c29f26c4ca6ac87406e8a39932
refs/heads/master
2023-04-12T10:10:25.928466
2021-05-10T17:55:12
2021-05-10T17:55:12
366,129,289
1
0
null
null
null
null
UTF-8
Python
false
false
3,041
py
import asyncio from aiohttp import web import json import os from services.exchange_rate_repository import ExchangeRateRepository routes = web.RouteTableDef() repo = ExchangeRateRepository(host="redis", port=6379) loop = asyncio.get_event_loop() @routes.get('/convert') async def handle_converion(request: web.Request): resp = web.Response(status=400) resp.content_type = "application/json" resp.text = json.dumps({'status': 'failed'}) params = ["from", "to", "amount"] all_params_present = all(param in request.query for param in params) if not all_params_present: return web.Response( status=400, content_type='application/json', text=json.dumps({ 'status': 'failed', 'message': 'not all parameters specified'}) ) qfrom = request.query['from'] qto = request.query['to'] qamount = num(request.query['amount']) if qamount is None: return web.Response( status=400, content_type='application/json', text=json.dumps({ 'status': 'failed', 'message': 'parameter "amount" must be 0 or 1'}) ) factor = repo.get_rates(qfrom, qto) if factor is None: return web.Response( status=404, content_type='application/json', text=json.dumps({ 'status': 'failed', 'message': 'unknown currency names'}) ) resp = web.Response(status=200) resp_obj = { 'status': 'success', 'amount': qamount * factor } resp.text = json.dumps(resp_obj) return resp @routes.post('/database') async def handle_converion(request: web.Request): if 'merge' not in request.query or num(request.query['merge'] is None): return web.Response( status=400, text=json.dumps({ 'status': 'failed', 'message': 'parameter "merge" is not present'} ) ) merge = int(request.query['merge']) body = await request.content.read() new_rates = {} try: new_rates = json.loads(body) except Exception: return web.Response( status=400, text=json.dumps({ 'status': 'failed', 'message': 'data must be in json format'} ) ) repo.merge_rates(new_rates, merge) return web.Response( status=200, text=json.dumps({ 'status': 'success', 'message': 'successfuly updated'} ) ) def num(s): res = try_parse_int(s) if res: return res try: return float(s) except ValueError: return None def try_parse_int(s): try: return int(s) except ValueError: return None def main(): port = int(os.environ['APP_PORT']) app = web.Application(loop=loop) app.add_routes(routes) web.run_app(app, port=port) if __name__ == "__main__": main()
[ "stepanurfu@gmail.com" ]
stepanurfu@gmail.com
d600fb5bea2fd528892cf24f66259ab17cd92938
98dbbcd09777a3071c4decefff8b19d477f225a1
/plot_functions.py
e532e7d0e6c7f802fb7e4a06c37470423110bd63
[]
no_license
ChristianMorup/RL_Pong
bbcc54f20f44be68ffb80b8257af067755a87476
8304df69ef0cabbb4b12b65a1b547fc32eb49b8d
refs/heads/main
2023-01-23T10:03:09.283135
2020-12-01T18:01:38
2020-12-01T18:01:38
313,404,719
0
0
null
null
null
null
UTF-8
Python
false
false
8,335
py
import numpy as np import matplotlib.pyplot as plt from file_util import FileUtil from modes_and_types import * apply_filters = True #TODO: ## Proportional to the distance ## Font and width of figures is too small ## Put labels on top of the figure ## Number of missed balls ## Invert the scores #reward_types = all_rewards #modes = all_modes modes = [Mode.BOTH] reward_types = [RewardType.TRACKING_PROPORTIONAL_UNIDIRECTIONAL, RewardType.TRACKING_PROPORTIONAL_UNIDIRECTIONAL_WEIGHTED] def moving_average(a, n=3): ret = np.cumsum(a, dtype=float) ret[n:] = ret[n:] - ret[:-n] return ret[n - 1:] / n def map_reward_type_to_title(reward): if reward == RewardType.TRACKING: title = 'Simple tracking ' elif reward == RewardType.HITTING: title = 'Hitting ' elif reward == RewardType.TRACKING_AND_HITTING: title = 'Simple tracking + Hitting ' elif reward == RewardType.TRACKING_PROPORTIONAL: title = 'Tracking \n(Proportional to the distance - Unidirectional) ' elif reward == RewardType.TRACKING_PROPORTIONAL_UNIDIRECTIONAL: title = 'Tracking \n(Proportional to the distance - Both directions) ' elif reward == RewardType.TRACKING_PROPORTIONAL_UNIDIRECTIONAL_WEIGHTED: title = 'Tracking \n(Weighted and proportional to the distance - Unidirectional)' elif reward == RewardType.TRACKING_STEADY_NERVES: title = 'Steady nerves' else: title = '' return 'Cumulative reward - ' + title, 'Score - ' + title def filter_by_picking_less_values(data): x = [] y = [] for i in range(len(data)): if i % 5 == 0: x = np.append(x, data[i]) y = np.append(y, i) return x, y def summarize_simulations(score_data, reward_data, n_simulations=15): score_sum = [] reward_sum = [] # lowest_reward = find_lowest_first_reward_score(reward_data, n_simulations) for column in range(1000): score_total = 0 reward_total = 0 for row in range(n_simulations): score_total += (100 - score_data[row][column]) reward_total += reward_data[row][column] #reward_total += normalize_reward_score(lowest_reward, reward_data[row][column]) new_score_value = score_total // n_simulations new_reward_value = reward_total // n_simulations score_sum = np.append(score_sum, new_score_value) reward_sum = np.append(reward_sum, new_reward_value) return filter(score_sum, reward_sum) def summarize_reward_simulations(reward_data, n_simulations=15): reward_sum = [] # lowest_reward = find_lowest_first_reward_score(reward_data, n_simulations) for column in range(1000): reward_total = 0 for row in range(n_simulations): reward_total += reward_data[row][column] # reward_total += normalize_reward_score(lowest_reward, reward_data[row][column]) new_reward_value = reward_total // n_simulations reward_sum = np.append(reward_sum, new_reward_value) reward_sum = moving_average(reward_sum, 3) reward_sum, reward_episodes = filter_by_picking_less_values(reward_sum) return reward_sum, reward_episodes def find_lowest_first_reward_score(data, n_simulations=15): lowest_reward = 0 for row in range(n_simulations): if lowest_reward > data[row][0]: lowest_reward = data[row][0] return lowest_reward def normalize_reward_score(lowest_reward, value): return value - lowest_reward def filter(score_sum, reward_sum): if apply_filters: score_sum = moving_average(score_sum, 3) reward_sum = moving_average(reward_sum, 3) score_sum, score_episodes = filter_by_picking_less_values(score_sum) reward_sum, reward_episodes = filter_by_picking_less_values(reward_sum) else: score_episodes = range(1000) reward_episodes = range(1000) return score_sum, score_episodes, reward_sum, reward_episodes def plot_simulations(folder_path, n_modes=3, n_simulations=10): for reward_type in reward_types: scores = [] rewards = [] pos_rewards = [] neg_rewards = [] episodes = [] for mode in modes: file_util = FileUtil(folder_path) s_data, r_data, r_pos_data, r_neg_data = file_util.read_files(reward_type, mode) s_sum, episodes, r_sum, r_episodes = summarize_simulations(s_data, r_data, n_simulations=n_simulations) scores = np.append(scores, s_sum) rewards = np.append(rewards, r_sum) if len(r_pos_data) > 0: r_pos_sum, pos_episodes = summarize_reward_simulations(r_pos_data, n_simulations=n_simulations) pos_rewards = np.append(pos_rewards, r_pos_sum) if len(r_neg_data) > 0: r_neg_sum, neg_episodes = summarize_reward_simulations(r_neg_data, n_simulations=n_simulations) neg_rewards = np.append(neg_rewards, r_neg_sum) if n_modes > 0: shape = (n_modes, len(episodes)) scores = scores.reshape(shape) rewards = rewards.reshape(shape) if len(pos_rewards) > 0: shape = (n_modes-1, len(episodes)) pos_rewards = pos_rewards.reshape(shape) if len(neg_rewards) > 0: shape = (n_modes-1, len(episodes)) neg_rewards = neg_rewards.reshape(shape) reward_title, score_title = map_reward_type_to_title(reward_type) # create_title(reward_type, mode) plt.clf() # plt.plot(episodes, scores, label="Both") plt.plot(episodes, scores[0], label="Both") if n_modes >= 2: plt.plot(episodes, scores[1], label="Negative") if n_modes >= 3: plt.plot(episodes, scores[2], label="Positive") plt.legend(loc='upper center', bbox_to_anchor=(0.5, 1.0), ncol=3, fancybox=True, shadow=False) plt.xlim(xmin=0, xmax=1000) plt.ylim(ymin=0, ymax=100) plt.ylabel('n Hits (max=100)') plt.xlabel('Episodes') plt.title(score_title) plt.show() plt.clf() # plt.plot(episodes, rewards, label="Both") plt.plot(episodes, rewards[0], label="Both") if n_modes >= 2: plt.plot(episodes, rewards[1], label="Negative") if n_modes >= 3: plt.plot(episodes, rewards[2], label="Positive") plt.legend(loc='upper center', bbox_to_anchor=(0.5, 1.0), ncol=3, fancybox=True, shadow=False) plt.xlim(xmin=0, xmax=1000) plt.ylabel('Cumulative reward') plt.xlabel('Episodes') plt.title(reward_title) plt.show() if len(pos_rewards) > 0: plt.clf() # plt.plot(episodes, rewards, label="Both") if n_modes - 1 >= 1: plt.plot(episodes, pos_rewards[0], label="Both") if n_modes - 1 >= 2: plt.plot(episodes, pos_rewards[1], label="Positive") plt.legend(loc='upper center', bbox_to_anchor=(0.5, 1.0), ncol=3, fancybox=True, shadow=False) plt.xlim(xmin=0, xmax=1000) plt.ylabel('Cumulative positive reward') plt.xlabel('Episodes') plt.title(reward_title) plt.show() if len(neg_rewards) > 0: plt.clf() # plt.plot(episodes, rewards, label="Both") if n_modes - 1 >= 1: plt.plot(episodes, neg_rewards[0], label="Both") if n_modes - 1 >= 2: plt.plot(episodes, neg_rewards[1].flatten(), label="Negative") plt.legend(loc='upper center', bbox_to_anchor=(0.5, 1.0), ncol=3, fancybox=True, shadow=False) plt.xlim(xmin=0, xmax=1000) plt.ylabel('Cumulative negative reward') plt.xlabel('Episodes') plt.title(reward_title) plt.show() def do_simple_plot(score): plt.clf() plt.plot(range(len(score)), score, label="0.1") plt.legend(bbox_to_anchor=(1.05, 1), loc='upper left', borderaxespad=0.) plt.xlim(xmin=0, xmax=400) plt.ylim(ymin=0) plt.ylabel('Number of \'back wall\'-hits') plt.xlabel('Episodes') plt.title('Test') plt.show() plot_simulations('sim_data/test/', n_modes=1, n_simulations=5)
[ "chrmorup@gmail.com" ]
chrmorup@gmail.com
5e070c29c4a7e73ae079e481606dc53c90ecce4e
32a60828c347544bffad94bf9716078c5717d760
/landing/apps/customuser/models.py
13cfc581f89337a9e3ba60cff0cd353c1ef9af35
[]
no_license
flikn/lazonaflikn
6c0b9c6eee85a0f23d1443af358ec8077b07b4f4
c8bfc2f50a03fc224f734ee2143280ea87a0964e
refs/heads/master
2020-08-11T17:19:49.092468
2015-02-16T22:35:41
2015-02-16T22:35:41
29,598,846
0
0
null
null
null
null
UTF-8
Python
false
false
2,727
py
from django.db import models from django.core import validators from django.contrib.auth.models import BaseUserManager, AbstractBaseUser from apps.subscribe.models import Subscriptor class MyUserManager(BaseUserManager): def create_user(self, email, password, username="", first_name="", last_name="", *args, **kwargs): """ Creates and saves a User. """ if not email: raise ValueError("Users must have an email address") user = self.model( email=self.normalize_email(email), username=username, first_name=first_name, last_name=last_name, raw_password=password, ) user.set_password(password) user.save(using=self._db) return user def create_superuser(self, email, password): """ Creates and saves a superuser with the given email and password. """ user = self.create_user( email, password=password, ) user.is_admin = True user.save(using=self._db) return user class MyUser(AbstractBaseUser): subscriptor = models.ForeignKey( Subscriptor, blank=True, null=True, ) email = models.EmailField( verbose_name="email address", max_length=50, unique=True, ) username = models.CharField( max_length=50, blank=True, ) first_name = models.CharField( max_length=50, blank=True, validators=[ validators.RegexValidator( r"^[a-zA-Z ]", "Invalid name." ), ], ) last_name = models.CharField( max_length=255, blank=True, validators=[ validators.RegexValidator( r"^[a-zA-Z ]", "Invalid name." ), ], ) is_admin = models.BooleanField( default=False, ) was_registered = models.BooleanField( default=False, ) was_subscribed = models.BooleanField( default=False, ) raw_password = models.CharField( max_length=100, blank=True, null=True, ) objects = MyUserManager() USERNAME_FIELD = "email" def get_full_name(self): return " ".join([self.first_name, self.last_name]) def get_short_name(self): return self.email def __str__(self): return self.email def has_perm(self, perm, obj=None): return True def has_module_perms(self, app_label): return True @property def is_staff(self): return self.is_admin from .signals import update_user
[ "oscar.gi.cast@gmail.com" ]
oscar.gi.cast@gmail.com
310ba0cb9368a175620ca3cbcbd62104bf3f9f8b
edc1f1369794a4a1c499c6e9d5fe49a712657611
/algorithms/leetcode_all/560.subarray-sum-equals-k/subarray-sum-equals-k.py
74c28e9996672f15fe435da46bf9edd7cf5ffdc2
[]
no_license
williamsyb/mycookbook
93d4aca1a539b506c8ed2797863de6da8a0ed70f
dd917b6eba48eef42f1086a54880bab6cd1fbf07
refs/heads/master
2023-03-07T04:16:18.384481
2020-11-11T14:36:54
2020-11-11T14:36:54
280,005,004
2
0
null
2023-03-07T02:07:46
2020-07-15T23:34:24
Python
UTF-8
Python
false
false
379
py
class Solution(object): def subarraySum(self, nums, k): """ :type nums: List[int] :type k: int :rtype: int """ preSum = ans = 0 visit = {0: 1} for i, n in enumerate(nums): preSum += n ans += visit.get(preSum - k, 0) visit[preSum] = visit.get(preSum, 0) + 1 return ans
[ "william_sun1990@hotmail.com" ]
william_sun1990@hotmail.com
11b9bf5a469cbefb5d55ecbc166fdf0b95d5e6a5
d2bb13cec7faf28e3d268312298f03c99806bd8b
/IPTS-16891-Dy2Ti2O7/norm_mesh_symm_All_rwp_100mK_7.py
d66eaa8f6dc140ec0ed3f53c2db9c0369b379c0f
[]
no_license
rosswhitfield/corelli
06a91c26556ea788f20f973a1018a56e82a8c09a
d9e47107e3272c4457aa0d2e0732fc0446f54279
refs/heads/master
2021-08-07T14:04:24.426151
2021-08-03T19:19:05
2021-08-03T19:19:05
51,771,543
0
0
null
null
null
null
UTF-8
Python
false
false
5,488
py
from mantid.simpleapi import * from mantid.geometry import SymmetryOperationFactory import numpy as np # about information on where the data are and where to save iptsfolder= "/SNS/CORELLI/IPTS-16891/" outputdir="/SNS/users/rwp/corelli/IPTS-16891-Dy2Ti2O7/" nxfiledir=iptsfolder + "nexus/" ccfiledir = iptsfolder +"shared/autoreduce/" UBfile = iptsfolder+"shared/DTO_UB_111Vertical.mat" reducedfile_prefix = "DTO_cc" LoadNexus(Filename='/SNS/CORELLI/shared/Vanadium/2016B/SolidAngle20160720NoCC.nxs', OutputWorkspace='sa') LoadNexus(Filename='/SNS/CORELLI/shared/Vanadium/2016B/Spectrum20160720NoCC.nxs', OutputWorkspace='flux') MaskBTP(Workspace='sa',Bank="1-30,62-91") MaskBTP(workspace='sa',Pixel='1-16,200-256') #Mask the magnet MaskBTP(Workspace='sa',Bank="49",Tube="1") MaskBTP(Workspace='sa',Bank="54",Tube="1") MaskBTP(Workspace='sa',Bank="58",Tube="13-16",Pixel="80-130") MaskBTP(Workspace='sa',Bank="59",Tube="1-4",Pixel="80-130") # Get UBs LoadEmptyInstrument(Filename='/SNS/CORELLI/shared/Calibration/CORELLI_Definition_cal_20160310.xml', OutputWorkspace='ub') LoadIsawUB(InputWorkspace='ub', Filename=UBfile) ub=mtd['ub'].sample().getOrientedLattice().getUB() print "Starting UB :" print ub #DTO Fd-3m (227) general position has 192 symmety operations. symOps = SymmetryOperationFactory.createSymOps(\ "x,y,z; -x,-y,z; -x,y,-z; x,-y,-z;\ z,x,y; z,-x,-y; -z,-x,y; -z,x,-y;\ y,z,x; -y,z,-x; y,-z,-x; -y,-z,x;\ y,x,-z; -y,-x,-z; y,-x,z; -y,x,z;\ x,z,-y; -x,z,y; -x,-z,-y; x,-z,y;\ z,y,-x; z,-y,x; -z,y,x; -z,-y,-x;\ -x,-y,-z; x,y,-z; x,-y,z; -x,y,z;\ -z,-x,-y; -z,x,y; z,x,-y; z,-x,y;\ -y,-z,-x; y,-z,x; -y,z,x; y,z,-x;\ -y,-x,z; y,x,z; -y,x,-z; y,-x,-z;\ -x,-z,y; x,-z,-y; x,z,y; -x,z,-y;\ -z,-y,x; -z,y,-x; z,-y,-x; z,y,x") ub_list=[] for sym in symOps: UBtrans = np.zeros((3,3)) UBtrans[0] = sym.transformHKL([1,0,0]) UBtrans[1] = sym.transformHKL([0,1,0]) UBtrans[2] = sym.transformHKL([0,0,1]) UBtrans=np.matrix(UBtrans.T) new_ub = ub*UBtrans print "Symmetry transform for "+sym.getIdentifier() print UBtrans print "New UB:" print new_ub ub_list.append(new_ub) #load in background #bkg=LoadEventNexus('/SNS/CORELLI/IPTS-15796/nexus/CORELLI_28124.nxs.h5') #bkg=LoadNexus('/SNS/CORELLI/IPTS-15796/shared/autoreduce/CORELLI_28124_elastic.nxs') #MaskDetectors(Workspace=bkg,MaskedWorkspace='sa') #pc_bkg=sum(bkg.getRun()['proton_charge'].value) #print 'pc_bkg=:'+str(pc_bkg) #T=1.8 K runs = range(34599,34635,1) #T=100 mK runs = range(34635,34653,1) totalrun = len(runs) print "Total number of runs %d" %totalrun if mtd.doesExist('normMD'): DeleteWorkspace('normMD') if mtd.doesExist('dataMD'): DeleteWorkspace('dataMD') #for r in runs: for index, r in enumerate(runs): print index, ' Processing run : %s' %r num=0 print 'Loading run number:'+ str(r) #filename='/SNS/CORELLI/IPTS-15526/nexus/CORELLI_'+str(r)+'.nxs.h5' #dataR=LoadEventNexus(Filename=filename) filename=ccfiledir+'CORELLI_'+str(r)+'_elastic.nxs' dataR=LoadNexus(Filename=filename) LoadInstrument(Workspace= dataR, Filename='/SNS/CORELLI/shared/Calibration/CORELLI_Definition_cal_20160310.xml',RewriteSpectraMap=False) MaskDetectors(Workspace=dataR,MaskedWorkspace='sa') pc_data=sum(dataR.getRun()['proton_charge'].value) print 'pc_data=:'+str(pc_data) #dataR=dataR - bkg*pc_data/pc_bkg # subtract the background if a background file was provided. Please make sure that the data were treated in the same way in terms of proton charge. if mtd.doesExist('Bkg'): bkg = mtd['Bkg'] ratio = pc_data/pc_bkg bkg_c = bkg*ratio Minus(LHSWorkspace=dataR, RHSWorkspace=bkg_c, OutputWorkspace=dataR) dataR=ConvertUnits(dataR,Target="Momentum",EMode="Elastic") dataR=CropWorkspace(dataR,XMin=2.5,XMax=10) SetGoniometer(dataR,Axis0="BL9:Mot:Sample:Axis2,0,1,0,1") LoadIsawUB(InputWorkspace=dataR,Filename=UBfile) for ub in ub_list: #for index, ub in enumerate(ub_list): #print "index, using UB ", (index+1), ":" num += 1 print "Run number"+str(r)+" Using UB:"+str(num) print ub SetUB(dataR, UB=ub) md=ConvertToMD(InputWorkspace=dataR,QDimensions='Q3D',dEAnalysisMode='Elastic', Q3DFrames='HKL', QConversionScales='HKL',MinValues='-7.1,-7.1,-7.1',MaxValues='7.1,7.1,7.1') a1,b1=MDNormSCD(InputWorkspace='md',FluxWorkspace='flux',SolidAngleWorkspace='sa', AlignedDim0="[H,0,0],-7.01,7.01,701", AlignedDim1="[0,K,0],-7.01,7.01,701", AlignedDim2="[0,0,L],-7.01,7.01,701") if mtd.doesExist('dataMD'): dataMD=dataMD+a1 else: dataMD=CloneMDWorkspace(a1) if mtd.doesExist('normMD'): normMD=normMD+b1 else: normMD=CloneMDWorkspace(b1) normData_CC=dataMD/normMD SaveMD('dataMD',Filename=outputdir+'DTO_datacc_48sym_Temp100mK_7.nxs') SaveMD('normMD',Filename=outputdir+'DTO_normcc_48sym_Temp100mK_7.nxs') SaveMD('normData_CC',Filename=outputdir+'DTO_normdatacc_48sym_Temp100mK_7.nxs') # group the data #data6K=GroupWorkspaces(datatoMerge) #md6K=GroupWorkspaces(mdtoMerge)
[ "whitfieldre@ornl.gov" ]
whitfieldre@ornl.gov
5959f52e07a61191622f22f7491aa2229167184c
ba104a6a6c3f84d2e96fb3c304af781d881bed8e
/bbs/dbproxy.py
1d37004c38872fff079b3090d675d41028b73a59
[]
no_license
jonny290/yos-x84
463a7f166ee59f2dd2a7a197352db1341dad8b66
7b03a35f12d2b7a10fa4709b09107935c6f14000
refs/heads/master
2021-01-20T06:59:22.869289
2015-01-17T06:13:32
2015-01-17T06:13:32
24,814,644
2
0
null
null
null
null
UTF-8
Python
false
false
4,902
py
""" Database proxy helper for X/84. """ import time class DBProxy(object): """ Provide dictionary-like object interface to a database. a database call, such as __len__() or keys() is issued as a command to the main engine, which spawns a thread to acquire a lock on the database and return the results via IPC pipe transfer. """ def __init__(self, schema, table='unnamed'): """ Arguments: schema: database key, to become basename of .sqlite3 files. """ self.schema = schema self.table = table def proxy_iter(self, method, *args): """ Iterable proxy for dictionary methods called over IPC pipe. """ import x84.bbs.session event = 'db=%s' % (self.schema,) session = x84.bbs.session.getsession() session.flush_event(event) session.send_event(event, (self.table, method, args)) data = session.read_event(event) assert data == (None, 'StartIteration'), ( 'iterable proxy used on non-iterable, %r' % (data,)) data = session.read_event(event) while data != (None, StopIteration): yield data data = session.read_event(event) session.flush_event(event) def proxy_method(self, method, *args): """ Proxy for dictionary methods called over IPC pipe. """ import x84.bbs.session event = 'db-%s' % (self.schema,) session = x84.bbs.session.getsession() session.send_event(event, (self.table, method, args)) return session.read_event(event) def acquire(self, blocking=True, stale=2.0): """ Acquire a fine-grained BBS-global lock, blocking or non-blocking. When invoked with the blocking argument set to True (the default), block until the lock is acquired, and return True. When invoked with the blocking argument set to False, do not block. Returns False if lock is not acquired. If the engine has held the lock longer than ``stale`` seconds, the lock is granted anyway. """ import x84.bbs.session event = 'lock-%s/%s' % (self.schema, self.table) session = x84.bbs.session.getsession() while True: session.send_event(event, ('acquire', stale)) data = session.read_event(event) if data is True or not blocking: return data time.sleep(0.1) def release(self): """ Release bbs-global lock on database. """ import x84.bbs.session event = 'lock-%s/%s' % (self.schema, self.table) session = x84.bbs.session.getsession() return session.send_event(event, ('release', None)) # pylint: disable=C0111 # Missing docstring def __contains__(self, key): return self.proxy_method('__contains__', key) __contains__.__doc__ = dict.__contains__.__doc__ def __getitem__(self, key): return self.proxy_method('__getitem__', key) __getitem__.__doc__ = dict.__getitem__.__doc__ def __setitem__(self, key, value): return self.proxy_method('__setitem__', key, value) __setitem__.__doc__ = dict.__setitem__.__doc__ def __delitem__(self, key): return self.proxy_method('__delitem__', key) __delitem__.__doc__ = dict.__delitem__.__doc__ def get(self, key, default=None): return self.proxy_method('get', key, default) get.__doc__ = dict.get.__doc__ def has_key(self, key): return self.proxy_method('has_key', key) has_key.__doc__ = dict.has_key.__doc__ def setdefault(self, key, value): return self.proxy_method('setdefault', key, value) setdefault.__doc__ = dict.setdefault.__doc__ def update(self, *args): return self.proxy_method('update', *args) update.__doc__ = dict.update.__doc__ def __len__(self): return self.proxy_method('__len__') __len__.__doc__ = dict.__len__.__doc__ def values(self): return self.proxy_method('values') values.__doc__ = dict.values.__doc__ def items(self): return self.proxy_method('items') items.__doc__ = dict.items.__doc__ def iteritems(self): return self.proxy_iter('iteritems') iteritems.__doc__ = dict.iteritems.__doc__ def iterkeys(self): return self.proxy_iter('iterkeys') iterkeys.__doc__ = dict.iterkeys.__doc__ def itervalues(self): return self.proxy_iter('itervalues') itervalues.__doc__ = dict.itervalues.__doc__ def keys(self): return self.proxy_method('keys') keys.__doc__ = dict.keys.__doc__ def pop(self): return self.proxy_method('pop') pop.__doc__ = dict.pop.__doc__ def popitem(self): return self.proxy_method('popitem') popitem.__doc__ = dict.popitem.__doc__
[ "root@5c2973472a80.(none)" ]
root@5c2973472a80.(none)
2da22543b0df8381720ed441872ed8dab2d2f559
4948250bd7ca92e47742bb40e89dfc74c3f50463
/projectwork/testapp/migrations/0007_auto_20210201_1657.py
60b45419969915a73692c4cfa10d98d26bb73a53
[]
no_license
sumitnicmar/PetCover
e145c3b5dddb946aa05feb23f4ad8e0c17c54707
a4d478d3e141b9c4e5c0050800aa37e90580fad9
refs/heads/main
2023-03-10T16:04:40.744012
2021-02-15T18:58:02
2021-02-15T18:58:02
339,175,398
0
0
null
null
null
null
UTF-8
Python
false
false
476
py
# Generated by Django 3.1.3 on 2021-02-01 11:27 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('testapp', '0006_auto_20210201_1633'), ] operations = [ migrations.AlterField( model_name='appointment', name='user', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='testapp.customer'), ), ]
[ "sumitsoft1993@gmail.com" ]
sumitsoft1993@gmail.com
355606e8e2db99254d740fee5aa6d2a68fae9623
08c4c7ed66ba9f2312b914b639c8e63d6a094552
/shoppingX/asgi.py
45071adf66eacba11b97fbf427fd7cbef0540f2e
[]
no_license
aman-ash/E---Commerce_Website_Project
c93ab4c314de1de87e98d693515999f5dc8fc969
4e5fd2e982b96d9bc111027a87b15381ecc7b687
refs/heads/master
2023-07-15T06:33:53.290197
2021-08-24T06:44:24
2021-08-24T06:44:24
382,203,247
0
0
null
null
null
null
UTF-8
Python
false
false
395
py
""" ASGI config for shoppingX project. It exposes the ASGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.2/howto/deployment/asgi/ """ import os from django.core.asgi import get_asgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'shoppingX.settings') application = get_asgi_application()
[ "amanmahore.ash@gmail.com" ]
amanmahore.ash@gmail.com
9ca0eebcc0fb73790d6d40f0aa355c00af6bea88
77f5208e62d4c222702184c920323029c15d37fe
/ml_project/work/features/nope_transformer.py
6efb522acc035a8202f9f84e79f6d47d4e768f6c
[]
no_license
made-ml-in-prod-2021/korowood
1f89a58b9126458d7f0d43068a0d28e6f230d11f
591cfd3341aae10618075123757011cf9465b751
refs/heads/main
2023-06-06T15:07:20.537003
2021-06-17T08:16:09
2021-06-17T08:16:09
354,288,201
0
0
null
2021-06-17T08:16:10
2021-04-03T12:37:46
Jupyter Notebook
UTF-8
Python
false
false
555
py
import numpy as np import pandas as pd from typing import List from sklearn.base import BaseEstimator, TransformerMixin class NopeTransformer(BaseEstimator, TransformerMixin): """ Dummy class to do pass-through dataset as is. """ def __init__(self, cols: List[str] = None, quantile: float = 0.05): self.cols = cols self.quantile = quantile def fit(self, X: pd.DataFrame, y: np.array = None) -> pd.DataFrame: return self def transform(self, X: pd.DataFrame, y: np.array = None) -> pd.DataFrame: return X
[ "yakorovka@gmail.com" ]
yakorovka@gmail.com
234c003b3b7516f29ad666cdb2a012b2e5e63f70
a90ce0aaa95b712c6b1a7c4fa584254f93d80451
/1-6-8.py
491c4a751d6c8962c237e878bf408bad937b28d0
[]
no_license
vvFell/stepic_course
a82743356144899d6e04ca51df67bbbaf501c91f
e8c662437b7ffb1abb7be4a8e08903184710b59f
refs/heads/master
2020-09-17T00:04:55.346737
2019-11-25T11:04:48
2019-11-25T11:04:48
223,927,087
0
0
null
null
null
null
UTF-8
Python
false
false
901
py
from selenium import webdriver import time link = "http://suninjuly.github.io/find_xpath_form" try: browser = webdriver.Chrome() browser.get(link) input1 = browser.find_element_by_tag_name('input') input1.send_keys("Ivan") input2 = browser.find_element_by_name('last_name') input2.send_keys("Petrov") input3 = browser.find_element_by_class_name('form-control.city') input3.send_keys("Smolensk") input4 = browser.find_element_by_id('country') input4.send_keys("Russia") button = browser.find_element_by_xpath("//button[@type='submit']") button.click() finally: # успеваем скопировать код за 30 секунд time.sleep(30) # закрываем браузер после всех манипуляций browser.quit() # не забываем оставить пустую строку в конце файла
[ "VSerov@profix.com" ]
VSerov@profix.com
3fc7db5e967c82b33be3e2e96772d7a1ad400835
2c3e61c057d00399e94985914556224fff6cc5be
/marketingFirm.py
077936d4cf5899dbe6303192c805fd836d3831aa
[]
no_license
wrightzachary/sweepstakes
7b09a7b41017cd870402bb639a6ad047ca314478
b5cc71a533d270b8cc9b6b23059a45dc798a9f18
refs/heads/main
2023-06-25T20:37:58.659130
2021-07-08T19:31:14
2021-07-08T19:31:14
379,325,841
0
0
null
null
null
null
UTF-8
Python
false
false
2,524
py
from sweepstakes import Sweepstakes from userInterface import UserInterface class MarketingFirm: def __init__(self): self.marketing_firm_name = "" self.sweepstakes_storage = ["Boat", "Weekend getaway"] # create sweepstakes def create_sweepstakes(self): chosen_name = UserInterface.get_user_input_string("\tEnter new sweepstakes name") UserInterface.display_message('') sweepstakes = Sweepstakes() self.sweepstakes_storage.append(chosen_name) print(f"\tYou created {chosen_name} ") UserInterface.display_message('') MarketingFirm.menu(self) # change firm name def change_marketing_firm_name(self): change_name = UserInterface.get_user_input_string("\tChange firm name:") new_name = change_name print(f'\tMarketing firm name changed to {new_name}') UserInterface.display_message('') self.marketing_firm_name = f'{new_name}' UserInterface.display_message('') # UserInterface.display_marketing_firm_menu_options(self.marketing_firm_name) MarketingFirm.menu(self) # select a sweepstake def select_sweepstakes(self): self.marketing_firm_name = UserInterface.get_user_input_string("\tSelect your desired sweepstake") UserInterface.display_message('') UserInterface.display_message(f'\tYou Selected {self.marketing_firm_name}') UserInterface.display_message('') UserInterface.display_sweepstakes_info('') response = int(input("\tPlease enter your selection: ")) if response == 1: UserInterface.display_marketing_firm_menu_options(self.marketing_firm_name) else: UserInterface.display_message("Not a valid selection") MarketingFirm.select_sweepstakes(self) # menu def menu(self): UserInterface.display_marketing_firm_menu_options(self.marketing_firm_name) response = int(input("\tPlease enter your selection: ")) UserInterface.display_message('') if response == 1: self.create_sweepstakes() elif response == 2: self.change_marketing_firm_name() elif response == 3: self.select_sweepstakes() elif response == 4: UserInterface.display_sweepstakes_menu_options(self.sweepstakes_storage) else: UserInterface.display_message("\tNot a valid selection") UserInterface.display_message('') MarketingFirm.menu(self)
[ "taylorzw96@gmail.com" ]
taylorzw96@gmail.com
df931b896dc092c4cf234d2b30a65601baa984cb
06d42cdf96dc12d3582ee2b107c481800eea95ae
/exercises/ex8_8.py
b57df9a6d5477ca300240ac35589a9f63c1a0c20
[]
no_license
GiaDieu/Python-exercises
236773137b1156dd9e5c2aef91b8e0ab4ecdb6fb
81ae81fc91d27c64f4f70922fb4c09cf8d73902d
refs/heads/master
2020-09-11T18:06:47.769509
2019-11-16T19:20:38
2019-11-16T19:20:38
222,147,357
1
0
null
null
null
null
UTF-8
Python
false
false
2,179
py
#!/usr/bin/env python3 __doc__ = ''' Viết script get_version nhận vào ngày ở format <month>/<day>/<year>. VD: 03/28/16 làm parameter và in ra một version được tính theo quy luật sau: - Version ở dạng format: <MAJOR>.<MINOR>.<PATCH>, vd: "6.9.2" - Từ ngày 09 tháng 02 năm 2016, phiên bản bắt đầu là "1.0.0" - Mỗi 28 ngày, MAJOR lại tăng thêm 1, MINOR và PATCH set về 0 - Mỗi tuần, MINOR tăng thêm 1 và PATCH sẽ set về 0 - Cứ mỗi ngày, PATCH lại tăng thêm 1. Yêu cầu: - Kiểm tra version thu được với lần lượt các input là "02/03/16", "09/06/16" với thởi điểm cuối là "06/23/17" Gợi ý: học viên sử dụng `sys.argv` hoặc module `argparse` ''' import sys import datetime def get_version(input_data): '''Trả về tên phiên bản như yêu cầu tại ``__doc__`` :param input_data: ngày format ở dạng <month>/<day>/<year>, ví dụ: "02/03/16" :rtype str: ''' # Sửa tên và function cho phù hợp, trả về kết quả yêu cầu. result = None start_time = datetime.datetime.strptime("02/09/16","%m/%d/%y") delta = datetime.datetime.strptime(input_data,"%m/%d/%y") - start_time days = delta.days MAJOR = days // 28 + 1 MINOR = (days % 28) // 7 PATCH = days % 7 result = "{}.{}.{}".format(MAJOR, MINOR, PATCH) return result def solve(input_data): '''Function `solve` dùng để `test`, học viên không cần chỉnh sửa gì thêm Chỉ thay đổi lại tên function của mình bên dưới cho phù hợp Gía trị trả về của hàm `solve` và `your_function` là như nhau :rtype str: ''' result = get_version(input_data) return result def main(): input_data = sys.argv[1] print(solve(input_data)) # sử dụng `sys.argv` hoặc `argparse` để gán gía trị yêu cầu # vào biến `input_data` # Xoá dòng sau và viết code vào đây set các giá trị phù hợp # raise NotImplementedError("Học viên chưa thực hiện truyền input_data") if __name__ == "__main__": main()
[ "giadieuly@gmail.com" ]
giadieuly@gmail.com
2622b01ab6dbae6e902dd8647177dd9c4dd6092a
64e899ae6199b7915585f7f903b2ec221aa0b8df
/common/modelanalysis.py
a0b3977391b8709375b3571f290d6da0707ec2f6
[]
no_license
sharma0611/dsworkflows
642f36dc0f7369c319eddbf27cd7bb734c9ab7a4
95dacb81581193d57fbc318ad1d9ef2c83b897cd
refs/heads/master
2021-09-15T01:38:11.201611
2018-05-23T14:31:52
2018-05-23T14:31:52
118,946,923
0
0
null
null
null
null
UTF-8
Python
false
false
13,940
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ **************************************** Shivam Sharma (CIH745) Data Science Co-op Summer 2017 **************************************** Module: Model Analysis Purpose: Hosts the functions that interact with sklearn-interface models to provide analysis on statistical metrics """ from common.utils import load_obj, save_obj import pandas as pd from numpy import array, random import numpy as np import os import re from sklearn.metrics import mean_squared_error, r2_score, confusion_matrix from importlib import import_module import importlib.util import matplotlib as mpl mpl.use("TkAgg") import matplotlib.backends.backend_pdf import matplotlib.pyplot as plt import matplotlib.gridspec as gridspec import seaborn as sns from common.univariateanalysis import apply_spec_to_df from common.transforms import inverse_dictionary #metrics def insert_random_var(seed, new_var, dataframe): """Function to insert random variable into Pandas DataFrame. """ random.seed(seed) dataframe[new_var] = random.random_sample(dataframe.shape[0]) #function used to select subset of another list, elements in order of first list def ordered_subset(preserve_order_list, cut_list): d = {k:v for v,k in enumerate(preserve_order_list)} new = set(preserve_order_list).intersection(cut_list) new = list(new) new.sort(key=d.get) return new def mse_r2_perf_output(actual_y, pred_y, title, fig=None, ax=None): if fig is None and ax is None: fig, ax = plt.subplots() elif fig is None: fig = ax.get_figure() elif ax is None: ax = fig.gca() mse = mean_squared_error(actual_y, pred_y) print(title + ":") print("MSE: %.4f" % mse) r2 = r2_score(actual_y, pred_y) print("r2: %.4f" % r2) ax.scatter(actual_y, pred_y, rasterized=True) ax.plot([actual_y.min(), actual_y.max()], [actual_y.min(), actual_y.max()], 'k--', lw=4) ax.set_xlabel('Actual') ax.set_ylabel('Predicted') ax.set_title(title) return mse, r2 def r2_mse_grab(test_y, test_y_pred): mse_test = mean_squared_error(test_y, test_y_pred) r2_test = r2_score(test_y, test_y_pred) return r2_test, mse_test def get_accuracy(actual_bands, predict_bands, interval): len_a = len(actual_bands) len_b = len(predict_bands) if len_a == len_b: accuracy = len([1 for x,y in zip(actual_bands, predict_bands) if x - interval <= y <= x + interval])/len_a return accuracy else: print("Different shapes between actual & predict of {0} and {1}".format(len_a, len_b)) exit(0) def cf_twentile_matrix(actual_twentile, predicted_twentile, title, normalize=True, fig=None, ax=None): if fig is None and ax is None: fig, ax = plt.subplots() elif fig is None: fig = ax.get_figure() elif ax is None: ax = fig.gca() cnf_matrix = confusion_matrix(actual_twentile, predicted_twentile) if normalize==True: cnf_matrix = cnf_matrix.astype('float') / np.amax(cnf_matrix) sns.heatmap(cnf_matrix, linewidths=.5, ax=ax) for i in range(4): accuracy = get_accuracy(actual_twentile, predicted_twentile, i) ax.text(s='{:.2%}'.format(accuracy) + " (x:x+-{0})".format(i), transform=ax.transAxes, x=0.75,y=0.95-i*0.07, fontsize=12) ax.set_title(title) ax.set_xlabel("Predicted") ax.set_ylabel("Actual") return fig, ax def tile_accuracy_hist(actual_tiles, predicted_tiles, tag="", predict_bands=True): cnf_matrix = confusion_matrix(actual_tiles, predicted_tiles) num_tiles = len(cnf_matrix) all_hist_figs = [] all_tiles = [x + 1 for x in range(num_tiles)] for i in range(num_tiles): if predict_bands: row = cnf_matrix[:,i] tiletype = "Predict" else: row = cnf_matrix[i] tiletype = "Actual" fig = plt.figure(figsize=(18, 6)) ax = fig.add_subplot(111) sns.set_style("darkgrid") ax.bar(left=all_tiles, height=row, tick_label=all_tiles, width=1) ax.set_title(tag + " Distribution for "+tiletype+" Tile " + str(i+1)) ax.set_ylabel("# of predictions") ax.set_xlabel("Predicted Tile") ax.set_xlim(0, num_tiles+1) #annotate each bar with percent of predictions total_predictions = sum(row) for p in ax.patches: ax.annotate('{:.2%}'.format(p.get_height()/total_predictions), (p.get_x() + (0 * p.get_width()), p.get_height() * 1.005), fontsize=10) all_hist_figs.append(fig) return all_hist_figs def mse_r2_graph(actual_y, pred_y, title, fig=None, ax=None): if fig is None and ax is None: fig, ax = plt.subplots() elif fig is None: fig = ax.get_figure() elif ax is None: ax = fig.gca() mse = mean_squared_error(actual_y, pred_y) print("MSE: %.4f" % mse) r2 = r2_score(actual_y, pred_y) print("r2: %.4f" % r2) ax.scatter(actual_y, pred_y) ax.plot([actual_y.min(), actual_y.max()], [actual_y.min(), actual_y.max()], 'k--', lw=4) ax.set_xlabel('Actual') ax.set_ylabel('Predicted') ax.set_title(title) return mse, r2 def predict_metrics(actual_y, predict_y, actual_bands, predict_bands, tag, fig, ax_array): #plot scatter plot of actual & predicted mse, r2 = mse_r2_perf_output(actual_y, predict_y, tag, fig, ax_array[0]) #create confusion matrix cf_twentile_matrix(actual_bands, predict_bands, tag, True, fig, ax_array[1]) mean_camaro = pd.DataFrame({'actual_y': actual_y, 'actual_twentiles': actual_bands, 'predicted_y': predict_y, 'predicted_twentiles': predict_bands }) actual = mean_camaro[['actual_twentiles', 'actual_y']].groupby(['actual_twentiles']).mean() pred = mean_camaro[['predicted_twentiles', 'predicted_y']].groupby(['predicted_twentiles']).mean() ax = actual.plot(ax=ax_array[2],rasterized=True) ax1 = pred.plot(ax=ax,rasterized=True) ax1.set_title(tag) ax1.set_ylabel("Mean predicted value for each tile") ax1.set_xlabel("Tile") return mse, r2 def r2_and_mse(clf, test_y, test_X, tag, estimators, fig=None, ax=None): if fig is None and ax is None: fig, ax = plt.subplots() elif fig is None: fig = ax.get_figure() elif ax is None: ax = fig.gca() test_r2 = np.zeros((estimators,), dtype=np.float64) test_mse = np.zeros((estimators,), dtype=np.float64) for i, y_pred in enumerate(clf.staged_predict(test_X)): test_r2[i] = r2_score(test_y , y_pred) test_mse[i] = mean_squared_error(test_y, y_pred) ax.plot(np.arange(estimators) + 1, test_r2, 'b-', label='R2', rasterized=True) ax.plot(np.arange(estimators) + 1, test_mse, 'r-', label='MSE', rasterized=True) ax.legend(loc='upper right') ax.set_xlabel('Boosting Iterations') ax.set_ylabel('R2 / MSE') ax.set_title(tag) return fig,ax def deviance(clf, test_y, test_X, tag, estimators, fig=None, ax=None): if fig is None and ax is None: fig, ax = plt.subplots() elif fig is None: fig = ax.get_figure() elif ax is None: ax = fig.gca() test_score = np.zeros((estimators,), dtype=np.float64) for i, y_pred in enumerate(clf.staged_predict(test_X)): test_score[i] = clf.loss_(test_y, y_pred) ax.set_title('Deviance') ax.plot(np.arange(estimators) + 1, clf.train_score_, 'b-', label='Training Set Deviance', rasterized=True) ax.plot(np.arange(estimators) + 1, test_score, 'r-', label='Test Set Deviance', rasterized=True) ax.legend(loc='upper right') ax.set_xlabel('Boosting Iterations') ax.set_ylabel('Deviance') return fig, ax import lightgbm as lgb def model_metrics_lgb(clf): fig3 = plt.figure(figsize=(8, 11)) gs3 = gridspec.GridSpec(2, 1) ax7 = fig3.add_subplot(gs3[0]) ax8 = fig3.add_subplot(gs3[1]) lgb.plot_metric(clf, metric="l2", ax=ax7, title="l2 during Training") lgb.plot_metric(clf, metric="huber", ax=ax8, title="Huber Loss during Training") gs3.tight_layout(fig3, rect=[0.05,0.05,0.95,0.95], pad=0.5) return [fig3] def model_metrics_sklearn(clf, estimators, actual_test_y, test_X, tag): fig3 = plt.figure(figsize=(8, 11)) gs3 = gridspec.GridSpec(2, 1) ax7 = fig3.add_subplot(gs3[0]) ax8 = fig3.add_subplot(gs3[1]) r2_and_mse(clf, actual_test_y, test_X, tag, estimators, fig3, ax7) deviance(clf, actual_test_y, test_X, tag, estimators, fig3, ax8) gs3.tight_layout(fig3, rect=[0.05,0.05,0.95,0.95], pad=0.5) return [fig3] #getting an r2 from model def r2_model(): #all output columns in dataframe y_pattern = ".*_" + str(y) + '$' r = re.compile(y_pattern) arr_y = filter(r.match, train_df.columns.tolist()) if features_to_use: if type(features_to_use) == str: features_to_use = eval(features_to_use) X = ordered_subset(X, features_to_use) train_X_df = train_df[X] train_X_arr = array(train_X_df) test_X_df = test_df[X] test_X_arr = array(test_X_df) #getting all y arrays y_label = transform + "_" + y train_y_series = train_df[y_label] train_y_arr = array(train_y_series) test_y_series = test_df[y_label] test_y_arr = array(test_y_series) #predict test_y_pred = model.predict(test_X_arr) train_y_pred = model.predict(train_X_arr) #get metrics r2_test, mse_test = r2_mse_grab(test_y_arr, test_y_pred) r2_train, mse_train = r2_mse_grab(train_y_arr, train_y_pred) #normalised mse by dividing by range of test_y_arr & train_y_arr respectively #nmse_test = mse_test / abs(max(test_y_arr) - min(test_y_arr)) #nmse_train = mse_train / abs(max(train_y_arr) - min(train_y_arr)) #print("test r2 {0:.2f}".format(r2_test)) #print("test mse {0:.2f}".format(mse_test)) #print("train r2 {0:.2f}".format(r2_train)) #print("train mse {0:.2f}".format(mse_train)) myseries = pd.Series([model_num, r2_test, mse_test, r2_train, mse_train]) myseries.index = ["ModelName", "test_r2", "test_mse", "train_r2", "train_mse"] return myseries def r2_compare(modeldb_path, impute_dir, y, exportpath=None, SpecialTag=None): tag = SpecialTag if os.path.isfile(modeldb_path): modeldb = load_obj(modeldb_path) else: print("modeldb not found") return cols = modeldb.columns.tolist() if "test_r2" not in cols: curr_db = modeldb elif tag: query = "r2_test > 0 | SpecialTag == " + str(tag) curr_db = modeldb.query(query) else: curr_db = modeldb.query("r2_test > 0") #load imputed data cooked_data_file = impute_dir + "/imputed.pk" train_fp = impute_dir + "/train.pk" test_fp = impute_dir + "/test.pk" cooked_df = load_obj(cooked_data_file) train_i = load_obj(train_fp) train_df = cooked_df.iloc[train_i] test_i = load_obj(test_fp) test_df = cooked_df.iloc[test_i] #get all metrics from DF temp_metrics_df = curr_db.apply(lambda row: r2_model(row["FullPath"],row["TransformTag"], y, row['ModelNum'], train_df, test_df), axis=1) new_columns = ['ModelNum', 'r2_test', 'mse_test', 'r2_train', 'mse_train'] temp_metrics_df.columns = new_columns modeldb = pd.merge(modeldb, temp_metrics_df, how='left', on='ModelNum') #save_obj(modeldb, modeldb_path) #CODE TO MAKE A CONFUSION MATRIX def train_test_confusion_plot_full(predicted_train, predicted_test, actual_train, actual_test, y, curr_tile, rev_transform_spec, full_dist=True): pred_train_df = pd.DataFrame({y: predicted_train}) pred_train_df = apply_spec_to_df(y, rev_transform_spec, pred_train_df) predicted_train_tile = pred_train_df[y].apply(curr_tile) pred_test_df = pd.DataFrame({y: predicted_test}) pred_test_df = apply_spec_to_df(y, rev_transform_spec, pred_test_df) predicted_test_tile = pred_test_df[y].apply(curr_tile) actual_train_df = pd.DataFrame({y: actual_train}) actual_train_df = apply_spec_to_df(y, rev_transform_spec, actual_train_df) actual_train_tile = actual_train_df[y].apply(curr_tile) actual_test_df = pd.DataFrame({y: actual_test}) actual_test_df = apply_spec_to_df(y, rev_transform_spec, actual_test_df) actual_test_tile = actual_test_df[y].apply(curr_tile) #setup figure for test items fig1 = plt.figure(figsize=(8, 11)) gs1 = gridspec.GridSpec(3, 1) ax1 = fig1.add_subplot(gs1[0]) ax2 = fig1.add_subplot(gs1[1]) ax3 = fig1.add_subplot(gs1[2]) ax_array1 = [ax1, ax2, ax3] test_mse, test_r2 = predict_metrics(actual_test, predicted_test, actual_test_tile.values, predicted_test_tile.values, y + ' - Test' , fig1, ax_array1) gs1.tight_layout(fig1, rect=[0.05,0.05,0.95,0.95], pad=0.5) #setup figure for train items fig2 = plt.figure(figsize=(8, 11)) gs2 = gridspec.GridSpec(3, 1) ax4 = fig2.add_subplot(gs2[0]) ax5 = fig2.add_subplot(gs2[1]) ax6 = fig2.add_subplot(gs2[2]) ax_array2 = [ax4, ax5, ax6] train_mse, train_r2 = predict_metrics(actual_train, predicted_train, actual_train_tile.values, predicted_train_tile.values, y + ' - Train' , fig2, ax_array2) gs2.tight_layout(fig2, rect=[0.05,0.05,0.95,0.95], pad=0.5) all_figs = [fig1, fig2] if full_dist: hist_figs = tile_accuracy_hist(actual_test_tile.values, predicted_test_tile.values, "Test", True) all_figs = all_figs + hist_figs return all_figs
[ "sharma.shivam0611@gmail.com" ]
sharma.shivam0611@gmail.com
26f2e95b46c06356ab3fa1f80b77aab1536140f8
67e1ccd069d6194158d32961c6a1f0fa49a0681f
/01_intro/recombination.py
6413cf48e0cd0ee4a97fad5584378a820d17259f
[ "Unlicense" ]
permissive
arantzardzm/software-art-text
a9a5e8e601876334049d484b2ef238ae233d47c5
2af98427892126ee00b1de12f7fd60b061386247
refs/heads/master
2021-07-21T16:17:12.191445
2017-10-31T09:20:59
2017-10-31T09:20:59
null
0
0
null
null
null
null
UTF-8
Python
false
false
917
py
import sys import random from time import sleep actor_adjectives = ['imperceptible', 'defiant', 'howling', 'subtle', 'eerie', 'cute'] actor = ['feline', 'cat', 'purr', 'claws', 'friend', 'crawler'] act = ['jumps', 'caresses', 'envelops', 'touches', 'reaches into', 'pets'] victim_adjectives = ['frightened', 'foolish', 'eager', 'loud', 'quiet', 'unconscious', 'wide-eyed'] victim = ['shadow', 'shiver', 'glimpse', 'crawl', 'sinner', 'saviour', 'hand', 'smile', 'cry'] while True: print "the %s %s %s the %s." % (random.choice(actor_adjectives), random.choice(actor), random.choice(act), random.choice(victim_adjectives)), conclusion = random.random() if conclusion > 0.8: print "there is a %s." % random.choice(victim) exit() elif conclusion > 0.6: print "there is no %s." % random.choice(victim) exit() else: print "" sys.stdout.flush() sleep(3)
[ "pierre.depaz@gmail.com" ]
pierre.depaz@gmail.com
abd7adc1822c7a3ded2bfbb351e303bc38039614
99a310f6bb6c7a6c728f1b3ae78054487372042d
/aoc2019/intcode/state_machine.py
b68372737c28f105cbb818391176e19138743da5
[]
no_license
jepebe/aoc2018
46ce6b46479a0faf2c2970413af14a071dcfdb79
4bf91b99bec4b59529533ef70f24bf6496bada99
refs/heads/master
2023-01-11T16:44:42.125394
2023-01-06T06:27:14
2023-01-06T06:27:14
159,912,721
1
0
null
null
null
null
UTF-8
Python
false
false
5,926
py
from collections import defaultdict def get_address(state_machine, parameter, write_mode=False): mode = state_machine['parameter_modes'][parameter] pos = state_machine['pos'] if mode == 0: addr = state_machine['instructions'][pos] elif mode == 1: if write_mode: print('Writing in immediate mode?') addr = pos elif mode == 2: addr = state_machine['instructions'][pos] relative_pos = state_machine['relative_pos'] addr = addr + relative_pos else: raise ('Unknown addressing mode %i for read' % mode) return addr def read(state_machine, parameter): addr = get_address(state_machine, parameter) state_machine['pos'] += 1 if addr >= len(state_machine['instructions']): return state_machine['memory'][addr] else: return state_machine['instructions'][addr] def write(state_machine, parameter, value): addr = get_address(state_machine, parameter, write_mode=True) state_machine['pos'] += 1 if addr >= len(state_machine['instructions']): state_machine['memory'][addr] = value else: state_machine['instructions'][addr] = value def add(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) write(state_machine, 2, a + b) def multiply(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) write(state_machine, 2, a * b) def get_input(state_machine): if len(state_machine['input']) == 0: state_machine['wait'] = True state_machine['pos'] -= 1 state_machine['instruction_count'] -= 1 else: data = state_machine['input'].pop(0) write(state_machine, 0, data) def output(state_machine): value = read(state_machine, 0) state_machine['output'].append(value) if state_machine['output_enabled']: print('Output from state machine %s' % value) def jump_if_true(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) if a != 0: state_machine['pos'] = b def jump_if_false(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) if a == 0: state_machine['pos'] = b def less_than(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) write(state_machine, 2, 1 if a < b else 0) def equals(state_machine): a = read(state_machine, 0) b = read(state_machine, 1) write(state_machine, 2, 1 if a == b else 0) def adjust_relative(state_machine): a = read(state_machine, 0) state_machine['relative_pos'] += a def halt(state_machine): state_machine['halt'] = True # print('Instruction count: %i' % state_machine['instruction_count']) def create_state_machine(instructions): return { 'instructions': list(instructions), 'backup_instructions': list(instructions), 'memory': defaultdict(int), 'operation': 0, 'parameter_modes': [0], 'pos': 0, 'relative_pos': 0, 'instruction_count': 0, 'input': [], 'output': [], 'last_output': None, 'output_enabled': False, 'opcodes': { 1: add, 2: multiply, 3: get_input, 4: output, 5: jump_if_true, 6: jump_if_false, 7: less_than, 8: equals, 9: adjust_relative, 99: halt }, 'halt': False, 'wait': False } def reset_state_machine(state_machine): state_machine['instructions'] = list(state_machine['backup_instructions']) state_machine['memory'] = defaultdict(int) state_machine['operation'] = 0 state_machine['parameter_modes'] = [0] state_machine['pos'] = 0 state_machine['relative_pos'] = 0 state_machine['instruction_count'] = 0 state_machine['input'] = [] state_machine['output'] = [] state_machine['last_output'] = None state_machine['output_enabled'] = False state_machine['halt'] = False state_machine['wait'] = False def parse(state_machine): pos = state_machine['pos'] opcode = state_machine['instructions'][pos] op = opcode % 100 p1 = ((opcode - op) // 100) % 10 p2 = ((opcode - op) // 1000) % 10 p3 = ((opcode - op) // 10000) % 10 state_machine['operation'] = state_machine['opcodes'][op] state_machine['parameter_modes'] = [p1, p2, p3] state_machine['pos'] += 1 def run_state_machine(state_machine): while not state_machine['halt'] and not state_machine['wait']: parse(state_machine) operation = state_machine['operation'] operation(state_machine) state_machine['instruction_count'] += 1 def add_input(state_machine, data): state_machine['input'].append(data) if state_machine['wait']: state_machine['wait'] = False def get_output(state_machine): if not has_output(state_machine): raise UserWarning('No output available!') state_machine['last_output'] = state_machine['output'][0] return state_machine['output'].pop(0) def has_output(state_machine): return len(state_machine['output']) > 0 def get_last_output(state_machine): return state_machine['last_output'] def flush_output(state_machine): while has_output(state_machine): get_output(state_machine) def load_instructions(filename): with open(filename) as f: instructions = f.readline().split(',') instructions = [int(x) for x in instructions] return instructions def load_state_machine(filename): instructions = load_instructions(filename) return create_state_machine(instructions) def is_running(state_machine): return not state_machine['halt'] def print_output(state_machine): import sys while has_output(state_machine): v = get_output(state_machine) sys.stdout.write(str(v) if v > 255 else chr(v))
[ "jepebe@users.noreply.github.com" ]
jepebe@users.noreply.github.com
c895843d8e49416fd67fe7412ef83108b18815d3
1e12dc07222de77f9344b8c260d7b848a76d8d5e
/word_cloud.py
70771c0f40ffd11a3a814b39c076130f1220d916
[]
no_license
ljyw17/douban_movie_analysis
4dd61f3a090189ccd411844cae929f12e81b005c
8d520f582b19f8d85add8ad72540a27db98bba60
refs/heads/master
2020-07-11T03:38:22.936473
2019-08-26T12:42:42
2019-08-26T12:42:42
204,436,858
3
2
null
null
null
null
UTF-8
Python
false
false
733
py
from wordcloud import WordCloud import PIL .Image as image import numpy as np import jieba, pandas def trans_CN(text): word_list = jieba.cut(text) # 分词后在单独个体之间加上空格 result = " ".join(word_list) return result if __name__=="__main__": file_path = r"Pure Hearts Into Chinese Showbiz.xlsx" df = pandas.read_excel(file_path, header=0) text = "" for i in range(len(df)): text += df["short"][i] text = trans_CN(text) wordcloud = WordCloud( font_path = r"C:\Windows\Fonts\msyh.ttc" ).generate(text) image_produce = wordcloud.to_image() image_produce.show() image_produce.save("{}.jpg".format(file_path.split(".")[0]))
[ "noreply@github.com" ]
noreply@github.com
c8d1f101c814d254035989eb2539d01b9c470f46
5b7413d0de8f2dc34df372489d233f59b12a76c6
/2.3.4.py
e4a37510affe30a6cde8198dd5b085c0fdf4c18e
[]
no_license
vekhnyk/stepik---auto-tests-course
cd1e9ce54f242a49b1b4a7b6fe9a25c99866dd07
3508794aa955c58c79fd31d1412faec5dbd2a2cc
refs/heads/master
2022-08-21T20:39:31.324410
2020-05-11T07:46:17
2020-05-11T07:46:17
262,959,713
0
0
null
null
null
null
UTF-8
Python
false
false
569
py
from selenium import webdriver from time import sleep import math link = 'http://suninjuly.github.io/alert_accept.html' try: browser = webdriver.Chrome() browser.get(link) browser.find_element_by_class_name('btn-primary').click() confirm = browser.switch_to.alert confirm.accept() x = browser.find_element_by_id('input_value').text rez = str(math.log(abs(12*math.sin(int(x))))) browser.find_element_by_id('answer').send_keys(rez) browser.find_element_by_class_name('btn-primary').click() finally: sleep(15) browser.quit()
[ "vekhnyk@gmail.com" ]
vekhnyk@gmail.com
62be2d8d4dbc67fc085acf0750f18d8471c8142f
f008fe1057b00d19d7812e0373669763c9992f42
/python/utils/hpatch.py
1047e32c3a097bb0f224c9141aa398b918144cee
[ "BSD-2-Clause" ]
permissive
baiyancheng20/hpatches-benchmark
68a265f5e23f7ceabce06635818b8c6a73c3e578
23c3a6dcf92bff7bda492363908cbc52b2533472
refs/heads/master
2021-01-20T00:52:08.069991
2017-04-22T19:02:21
2017-04-22T19:02:21
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,448
py
import cv2 import numpy as np from glob import glob from joblib import Parallel, delayed import multiprocessing import pandas as pd import json import os import time import scipy import copy # all types of patches tps = ['ref','e1','e2','e3','e4','e5','h1','h2','h3','h4','h5',\ 't1','t2','t3','t4','t5'] def vis_patches(seq,tp,ids): """Visualises a set of types and indices for a sequence""" w = len(tp)*65 vis = np.empty((0, w)) # add the first line with the patch type names vis_tmp = np.empty((35, 0)) for t in tp: tp_patch = 255*np.ones((35,65)) cv2.putText(tp_patch,t,(5,25),cv2.FONT_HERSHEY_DUPLEX , 1,0,1) vis_tmp = np.hstack((vis_tmp,tp_patch)) vis = np.vstack((vis,vis_tmp)) # add the actual patches for idx in ids: vis_tmp = np.empty((65, 0)) for t in tp: vis_tmp = np.hstack((vis_tmp,get_patch(seq,t,idx))) vis = np.vstack((vis,vis_tmp)) return vis def get_patch(seq,t,idx): """Gets a patch from a sequence with type=t and id=idx""" return getattr(seq, t)[idx] def get_im(seq,t):#rename this as a general method """Gets a patch from a sequence with type=t and id=idx""" return getattr(seq, t) def load_splits(f_splits): """Loads the json encoded splits""" with open(f_splits) as f: splits = json.load(f) return splits def load_descrs(path,dist='L2',descr_type='',sep=','): """Loads *all* saved patch descriptors from a root folder""" print('>> Please wait, loading the descriptor files...') # get all folders in the descriptor root folder, except the 1st which is '.' t = [x[0] for x in os.walk(path)][1::] try: len(t) == 116 except: print("%r does not seem like a valid HPatches descriptor root folder." % (path)) seqs_l = Parallel(n_jobs=multiprocessing.cpu_count())\ (delayed(hpatch_descr)(f,name,descr_type,sep) for f in t) seqs = dict((l.name, l) for l in seqs_l) seqs['distance'] = dist seqs['dim'] = seqs_l[0].dim print('>> Descriptor files loaded.') return seqs ############### # PCA Methods # ############### # TODO add error if no training set - cant do PCA on test. (e.g. the # full/view/illum split) def compute_pcapl(descr,split): X = np.empty((0,descr['dim'])) for seq in split['train']: X = np.vstack((X,get_im(descr[seq],'ref'))) X -= np.mean(X, axis=0) Xcov = np.dot(X.T,X) Xcov = (Xcov + Xcov.T) / (2 * X.shape[0]); d, V = np.linalg.eigh(Xcov) vv = np.sort(d) cl = vv[int(0.6*len(vv))] d[d<=cl]=cl D = np.diag(1. / np.sqrt(d)) W = np.dot(np.dot(V, D), V.T) for seq in split['test']: print(seq) for t in tps: X = get_im(descr[seq],t) X -= np.mean(X, axis=0) X_pca = np.dot(X,W) X_pcapl = np.sign(X_pca) * np.power(np.abs(X_pca),0.5) norms = np.linalg.norm(X_pcapl,axis=1) X_proj = (X_pcapl.T / norms).T X_proj = np.nan_to_num(X_proj) setattr(descr[seq], t, X_proj) ################################ # Patch and descriptor classes # ################################ class hpatch_descr: """Class for loading an HPatches descriptor result .csv file""" itr = tps def __init__(self,base,name,descr_type='',sep=','): self.base = base self.name = base.split("/")[-1] for t in self.itr: descr_path = os.path.join(base, t+'.csv') df = pd.read_csv(descr_path,header=None,sep=sep).as_matrix() df = df.astype(np.float32) if descr_type=="bin_packed": df = df.astype(np.uint8) df = np.unpackbits(df, axis=1) setattr(self, t, df) self.N = df.shape[0] self.dim = df.shape[1] assert self.dim != 1, \ "Problem loading the .csv files. Please check the delimiter." class hpatch_sequence: """Class for loading an HPatches sequence from a sequence folder""" itr = tps def __init__(self,base): name = base.split('/') self.name = name[-1] self.base = base for t in self.itr: im_path = os.path.join(base, t+'.png') im = cv2.imread(im_path,0) self.N = im.shape[0]/65 setattr(self, t, np.split(im, self.N))
[ "v.balntas@imperial.ac.uk" ]
v.balntas@imperial.ac.uk
78e5bb79876dae249b0a9fd6b8d331c441ff4a64
f00f97b6450a39ad148eec02609b059c244a2e9c
/meaning of life.py
26a471413d33aac06d6f1109d17895d5bc102dcc
[]
no_license
BridgeFour4/pythonguiexamples
0eaae8fe71268161995af8269f40eab39cc719ed
79b7304ab2d207b215eb48064aa1b309ec169b6a
refs/heads/master
2020-04-27T14:43:32.889899
2019-03-11T20:11:17
2019-03-11T20:11:17
174,418,733
0
0
null
null
null
null
UTF-8
Python
false
false
1,362
py
from tkinter import * class Application(Frame): def __init__(self,master): super(Application,self).__init__(master) self.grid() self.create_widgets() def create_widgets(self): self.configure(bg="silver") self.labelone = Label(self,text= "Enter password for the secret life") self.labelone.grid(row=0, column=0,columnspan=2,sticky=EW) self.labeltwo= Label(self, text="password:") self.labeltwo.grid(row=1, column=0, sticky=EW) self.pw_ent =Entry(self,bg="black") self.pw_ent.grid(row=2, column=0, columnspan=2, sticky=W) self.buttonsubmit = Button(self, text="Submit", command=self.reveal) self.buttonsubmit.grid(row=2, column=1,columnspan=2, sticky=EW) self.secret_txt = Text(self, width=35, height=5, wrap=WORD,bg='orange',fg="white") self.secret_txt.grid(row=4, column=0, columnspan=3, sticky=E) def reveal(self): contents = self.pw_ent.get() if contents =="secret": message="42" else: message="That's not the correct password so I can't share the secret with you" self.secret_txt.delete(0.0, END) self.secret_txt.insert(0.0, message) root=Tk() root.title("Click Counter") root.geometry("300x200") root.configure(bg="black") app=Application(root) root.mainloop()
[ "nathan.broadbent@tooeleschools.org" ]
nathan.broadbent@tooeleschools.org
0aa4fad95e735af119da27c643164c508715fe23
58c122786263edf8aec4a6b6b4986b2f3d4ff1d5
/modules/s3/pyvttbl/qsturng.py
49e432d4a21df43c96c06038b370da50df36ded8
[ "MIT", "LicenseRef-scancode-unknown-license-reference" ]
permissive
andygimma/eden
c9d0819e87ef05ff607cac6120dbddc86e55bc31
716d5e11ec0030493b582fa67d6f1c35de0af50d
refs/heads/master
2021-01-15T21:54:03.240072
2012-11-16T05:13:11
2012-11-16T05:13:11
6,726,106
1
0
null
null
null
null
UTF-8
Python
false
false
47,225
py
# Copyright (c) 2011, Roger Lew [see LICENSE.txt] # This software is funded in part by NIH Grant P20 RR016454. """ Implementation of Gleason's (1999) non-iterative upper quantile studentized range approximation. According to Gleason this method should be more accurate than the AS190 FORTRAN algorithm of Lund and Lund (1983) and works from .5 <= p <= .999 (The AS190 only works from .9 <= p <= .99). It is more efficient then the Copenhaver & Holland (1988) algorithm (used by the _qtukey_ R function) although it requires storing the A table in memory. (q distribution) approximations in Python. see: Gleason, J. R. (1999). An accurate, non-iterative approximation for studentized range quantiles. Computational Statistics & Data Analysis, (31), 147-158. Gleason, J. R. (1998). A table of quantile points of the Studentized range distribution. http://www.stata.com/stb/stb46/dm64/sturng.pdf """ import math import scipy.stats import numpy as np inf = float('inf') __version__ = '0.2.1' # changelog # 0.1 - initial release # 0.1.1 - vectorized # 0.2 - psturng added # 0.2.1 - T, R generation script relegated to make_tbls.py # Gleason's table was derived using least square estimation on the tabled # r values for combinations of p and v. In total there are 206 # estimates over p-values of .5, .75, .9, .95, .975, .99, .995, # and .999, and over v (degrees of freedom) of (1) - 20, 24, 30, 40, # 60, 120, and inf. combinations with p < .95 don't have coefficients # for v = 1. Hence the parentheses. These coefficients allow us to # form f-hat. f-hat with the inverse t transform of tinv(p,v) yields # a fairly accurate estimate of the studentized range distribution # across a wide range of values. According to Gleason this method # should be more accurate than algorithm AS190 of Lund and Lund (1983) # and work across a wider range of values (The AS190 only works # from .9 <= p <= .99). R's qtukey algorithm was used to add tables # at .675, .8, and .85. These aid approximations when p < .9. # # The code that generated this table is called make_tbls.py and is # located in version control. A = {(0.1, 2.0): [-2.2485085243379075, -1.5641014278923464, 0.55942294426816752, -0.060006608853883377], (0.1, 3.0): [-2.2061105943901564, -1.8415406600571855, 0.61880788039834955, -0.062217093661209831], (0.1, 4.0): [-2.1686691786678178, -2.008196172372553, 0.65010084431947401, -0.06289005500114471], (0.1, 5.0): [-2.145077200277393, -2.112454843879346, 0.66701240582821342, -0.062993502233654797], (0.1, 6.0): [-2.0896098049743155, -2.2400004934286497, 0.70088523391700142, -0.065907568563272748], (0.1, 7.0): [-2.0689296655661584, -2.3078445479584873, 0.71577374609418909, -0.067081034249350552], (0.1, 8.0): [-2.0064956480711262, -2.437400413087452, 0.76297532367415266, -0.072805518121505458], (0.1, 9.0): [-2.3269477513436061, -2.0469494712773089, 0.60662518717720593, -0.054887108437009016], (0.1, 10.0): [-2.514024350177229, -1.8261187841127482, 0.51674358077906746, -0.044590425150963633], (0.1, 11.0): [-2.5130181309130828, -1.8371718595995694, 0.51336701694862252, -0.043761825829092445], (0.1, 12.0): [-2.5203508109278823, -1.8355687130611862, 0.5063486549107169, -0.042646205063108261], (0.1, 13.0): [-2.5142536438310477, -1.8496969402776282, 0.50616991367764153, -0.042378379905665363], (0.1, 14.0): [-2.3924634153781352, -2.013859173066078, 0.56421893251638688, -0.048716888109540266], (0.1, 15.0): [-2.3573552940582574, -2.0576676976224362, 0.57424068771143233, -0.049367487649225841], (0.1, 16.0): [-2.3046427483044871, -2.1295959138627993, 0.59778272657680553, -0.051864829216301617], (0.1, 17.0): [-2.2230551072316125, -2.2472837435427127, 0.64255758243215211, -0.057186665209197643], (0.1, 18.0): [-2.3912859179716897, -2.0350604070641269, 0.55924788749333332, -0.047729331835226464], (0.1, 19.0): [-2.4169773092220623, -2.0048217969339146, 0.54493039319748915, -0.045991241346224065], (0.1, 20.0): [-2.4264087194660751, -1.9916614057049267, 0.53583555139648154, -0.04463049934517662], (0.1, 24.0): [-2.3969903132061869, -2.0252941869225345, 0.53428382141200137, -0.043116495567779786], (0.1, 30.0): [-2.2509922780354623, -2.2309248956124894, 0.60748041324937263, -0.051427415888817322], (0.1, 40.0): [-2.1310090183854946, -2.3908466074610564, 0.65844375382323217, -0.05676653804036895], (0.1, 60.0): [-1.9240060179027036, -2.6685751031012233, 0.75678826647453024, -0.067938584352398995], (0.1, 120.0): [-1.9814895487030182, -2.5962051736978373, 0.71793969041292693, -0.063126863201511618], (0.1, inf): [-1.913410267066703, -2.6947367328724732, 0.74742335122750592, -0.06660897234304515], (0.5, 2.0): [-0.88295935738770648, -0.1083576698911433, 0.035214966839394388, -0.0028576288978276461], (0.5, 3.0): [-0.89085829205846834, -0.10255696422201063, 0.033613638666631696, -0.0027101699918520737], (0.5, 4.0): [-0.89627345339338116, -0.099072524607668286, 0.032657774808907684, -0.0026219007698204916], (0.5, 5.0): [-0.89959145511941052, -0.097272836582026817, 0.032236187675182958, -0.0025911555217019663], (0.5, 6.0): [-0.89959428735702474, -0.098176292411106647, 0.032590766960226995, -0.0026319890073613164], (0.5, 7.0): [-0.90131491102863937, -0.097135907620296544, 0.032304124993269533, -0.0026057965808244125], (0.5, 8.0): [-0.90292500599432901, -0.096047500971337962, 0.032030946615574568, -0.0025848748659053891], (0.5, 9.0): [-0.90385598607803697, -0.095390771554571888, 0.031832651111105899, -0.0025656060219315991], (0.5, 10.0): [-0.90562524936125388, -0.093954488089771915, 0.031414451048323286, -0.0025257834705432031], (0.5, 11.0): [-0.90420347371173826, -0.095851656370277288, 0.0321150356209743, -0.0026055056400093451], (0.5, 12.0): [-0.90585973471757664, -0.094449306296728028, 0.031705945923210958, -0.0025673330195780191], (0.5, 13.0): [-0.90555437067293054, -0.094792991050780248, 0.031826594964571089, -0.0025807109129488545], (0.5, 14.0): [-0.90652756604388762, -0.093792156994564738, 0.031468966328889042, -0.0025395175361083741], (0.5, 15.0): [-0.90642323700400085, -0.094173017520487984, 0.031657517378893905, -0.0025659271829033877], (0.5, 16.0): [-0.90716338636685234, -0.093785178083820434, 0.031630091949657997, -0.0025701459247416637], (0.5, 17.0): [-0.90790133816769714, -0.093001147638638884, 0.031376863944487084, -0.002545143621663892], (0.5, 18.0): [-0.9077432927051563, -0.093343516378180599, 0.031518139662395313, -0.0025613906133277178], (0.5, 19.0): [-0.90789499456490286, -0.09316964789456067, 0.031440782366342901, -0.0025498353345867453], (0.5, 20.0): [-0.90842707861030725, -0.092696016476608592, 0.031296040311388329, -0.0025346963982742186], (0.5, 24.0): [-0.9083281347135469, -0.092959308144970776, 0.031464063190077093, -0.0025611384271086285], (0.5, 30.0): [-0.90857624050016828, -0.093043139391980514, 0.031578791729341332, -0.0025766595412777147], (0.5, 40.0): [-0.91034085045438684, -0.091978035738914568, 0.031451631000052639, -0.0025791418103733297], (0.5, 60.0): [-0.91084356681030032, -0.091452675572423425, 0.031333147984820044, -0.0025669786958144843], (0.5, 120.0): [-0.90963649561463833, -0.093414563261352349, 0.032215602703677425, -0.0026704024780441257], (0.5, inf): [-0.91077157500981665, -0.092899220350334571, 0.032230422399363315, -0.0026696941964372916], (0.675, 2.0): [-0.67231521026565144, -0.097083624030663451, 0.027991378901661649, -0.0021425184069845558], (0.675, 3.0): [-0.65661724764645824, -0.08147195494632696, 0.02345732427073333, -0.0017448570400999351], (0.675, 4.0): [-0.65045677697461124, -0.071419073399450431, 0.020741962576852499, -0.0015171262565892491], (0.675, 5.0): [-0.64718875357808325, -0.064720611425218344, 0.019053450246546449, -0.0013836232986228711], (0.675, 6.0): [-0.64523003702018655, -0.059926313672731824, 0.017918997181483924, -0.0012992250285556828], (0.675, 7.0): [-0.64403313148478836, -0.056248191513784476, 0.017091446791293721, -0.0012406558789511822], (0.675, 8.0): [-0.64325095865764359, -0.053352543126426684, 0.016471879286491072, -0.0011991839050964099], (0.675, 9.0): [-0.64271152754911653, -0.051023769620449078, 0.01599799600547195, -0.0011693637984597086], (0.675, 10.0): [-0.64232244408502626, -0.049118327462884373, 0.015629704966568955, -0.0011477775513952285], (0.675, 11.0): [-0.64203897854353564, -0.047524627960277892, 0.015334801262767227, -0.0011315057284007177], (0.675, 12.0): [-0.64180344973512771, -0.046205907576003291, 0.015108290595438166, -0.0011207364514518488], (0.675, 13.0): [-0.64162086456823342, -0.045076099336874231, 0.0149226565346125, -0.0011126140690497352], (0.675, 14.0): [-0.64146906480198984, -0.044108523550512715, 0.014772954218646743, -0.0011069708562369386], (0.675, 15.0): [-0.64133915151966603, -0.043273370927039825, 0.014651691599222836, -0.0011032216539514398], (0.675, 16.0): [-0.64123237842752079, -0.042538925012463868, 0.014549992487506169, -0.0011005633864334021], (0.675, 17.0): [-0.64113034037536609, -0.041905699463005854, 0.014470805560767184, -0.0010995286436738471], (0.675, 18.0): [-0.64104137391561256, -0.041343885546229336, 0.014404563657113593, -0.0010991304223377683], (0.675, 19.0): [-0.64096064882827297, -0.04084569291139839, 0.014350159655133801, -0.0010993656711121901], (0.675, 20.0): [-0.64088647405089572, -0.040402175957178085, 0.014305769823654429, -0.0011001304776712105], (0.675, 24.0): [-0.64063763965937837, -0.039034716348048545, 0.014196703837251648, -0.0011061961945598175], (0.675, 30.0): [-0.64034987716294889, -0.037749651156941719, 0.014147040999127263, -0.0011188251352919833], (0.675, 40.0): [-0.6399990514713938, -0.036583307574857803, 0.014172070700846548, -0.0011391004138624943], (0.675, 60.0): [-0.63955586202430248, -0.035576938958184395, 0.014287299153378865, -0.0011675811805794236], (0.675, 120.0): [-0.63899242674778622, -0.034763757512388853, 0.014500726912982405, -0.0012028491454427466], (0.675, inf): [-0.63832682579247613, -0.034101476695520404, 0.014780921043580184, -0.0012366204114216408], (0.75, 2.0): [-0.60684073638504454, -0.096375192078057031, 0.026567529471304554, -0.0019963228971914488], (0.75, 3.0): [-0.57986144519102656, -0.078570292718034881, 0.021280637925009449, -0.0015329306898533772], (0.75, 4.0): [-0.56820771686193594, -0.0668113563896649, 0.018065284051059189, -0.0012641485481533648], (0.75, 5.0): [-0.56175292435740221, -0.058864526929603825, 0.016046735025708799, -0.0011052560286524044], (0.75, 6.0): [-0.55773449282066356, -0.053136923269827351, 0.014684258167069347, -0.0010042826823561605], (0.75, 7.0): [-0.55509524598867332, -0.048752649191139405, 0.013696566605823626, -0.00093482210003133898], (0.75, 8.0): [-0.55324993686191515, -0.045305558708724644, 0.012959681992062138, -0.00088583541601696021], (0.75, 9.0): [-0.55189259054026196, -0.042539819902381634, 0.012398791106424769, -0.00085083962241435827], (0.75, 10.0): [-0.55085384656956893, -0.040281425755686585, 0.01196442242722482, -0.00082560322161492677], (0.75, 11.0): [-0.55003198103541273, -0.038410176100193948, 0.011623294239447784, -0.00080732975034320073], (0.75, 12.0): [-0.54936541596319177, -0.036838543267887103, 0.011351822637895701, -0.0007940703654926442], (0.75, 13.0): [-0.54881015972753833, -0.035506710625568455, 0.011134691307865171, -0.0007846360016355809], (0.75, 14.0): [-0.54834094346071949, -0.034364790609906569, 0.010958873929274728, -0.00077796645357008291], (0.75, 15.0): [-0.54793602418304255, -0.033379237455748029, 0.010816140998057593, -0.00077344175064785099], (0.75, 16.0): [-0.54758347689728037, -0.032520569145898917, 0.010699240399358219, -0.00077050847328596678], (0.75, 17.0): [-0.54727115963795303, -0.031769277192927527, 0.010603749751170481, -0.0007688642392748113], (0.75, 18.0): [-0.54699351808826535, -0.031105476267880995, 0.010524669113016114, -0.00076810656837464093], (0.75, 19.0): [-0.54674357626419079, -0.030516967201954001, 0.010459478822937069, -0.00076808652582440037], (0.75, 20.0): [-0.54651728378950126, -0.029992319199769232, 0.010405694998386575, -0.0007686417223966138], (0.75, 24.0): [-0.54578309546828363, -0.028372628574010936, 0.010269939602271542, -0.00077427370647261838], (0.75, 30.0): [-0.54501246434397554, -0.026834887880579802, 0.010195603314317611, -0.00078648615954105515], (0.75, 40.0): [-0.54418127442022624, -0.025413224488871379, 0.010196455193836855, -0.00080610785749523739], (0.75, 60.0): [-0.543265189207915, -0.024141961069146383, 0.010285001019536088, -0.00083332193364294587], (0.75, 120.0): [-0.54224757817994806, -0.023039071833948214, 0.010463365295636302, -0.00086612828539477918], (0.75, inf): [-0.54114579815367159, -0.02206592527426093, 0.01070374099737127, -0.00089726564005122183], (0.8, 2.0): [-0.56895274046831146, -0.096326255190541957, 0.025815915364208686, -0.0019136561019354845], (0.8, 3.0): [-0.5336038380862278, -0.077585191014876181, 0.020184759265389905, -0.0014242746007323785], (0.8, 4.0): [-0.51780274285934258, -0.064987738443608709, 0.016713309796866204, -0.001135379856633562], (0.8, 5.0): [-0.50894361222268403, -0.056379186603362705, 0.014511270339773345, -0.00096225604117493205], (0.8, 6.0): [-0.50335153028630408, -0.050168860294790812, 0.01302807093593626, -0.00085269812692536306], (0.8, 7.0): [-0.49960934380896432, -0.045417333787806033, 0.011955593330247398, -0.00077759605604250882], (0.8, 8.0): [-0.49694518248979763, -0.041689151516021969, 0.011158986677273709, -0.00072497430103953366], (0.8, 9.0): [-0.4949559974898507, -0.038702217132906024, 0.010554360004521268, -0.0006875213117164109], (0.8, 10.0): [-0.49341407910162483, -0.036266788741325398, 0.010087354421936092, -0.00066060835062865602], (0.8, 11.0): [-0.49218129312493897, -0.034252403643273498, 0.0097218584838579536, -0.00064123459335201907], (0.8, 12.0): [-0.49117223957112183, -0.032563269730499021, 0.0094318583096021404, -0.00062725253852419032], (0.8, 13.0): [-0.49032781145131277, -0.031132495018324432, 0.0091999762562792898, -0.0006172944366003854], (0.8, 14.0): [-0.48961049628464259, -0.029906921170494854, 0.009012451847823854, -0.00061026211968669543], (0.8, 15.0): [-0.48899069793054922, -0.028849609914548158, 0.0088602820002619594, -0.00060548991575179055], (0.8, 16.0): [-0.48844921216636505, -0.027929790075266154, 0.00873599263877896, -0.00060242119796859379], (0.8, 17.0): [-0.48797119683309537, -0.027123634910159868, 0.0086338139869481887, -0.00060061821593399998], (0.8, 18.0): [-0.48754596864745836, -0.026411968723496961, 0.0085493196604705755, -0.00059977083160833624], (0.8, 19.0): [-0.48716341805691843, -0.025781422230819986, 0.0084796655915025769, -0.00059970031758323466], (0.8, 20.0): [-0.48681739197185547, -0.025219629852198749, 0.0084221844254287765, -0.00060023212822886711], (0.8, 24.0): [-0.48570639629281365, -0.023480608772518948, 0.008274490561114187, -0.000605681105792215], (0.8, 30.0): [-0.48455867067770253, -0.021824655071720423, 0.0081888502974720567, -0.00061762126933785633], (0.8, 40.0): [-0.48335478729267423, -0.020279958998363389, 0.0081765095914194709, -0.00063657117129829635], (0.8, 60.0): [-0.48207351944996679, -0.018875344346672228, 0.0082473997191472338, -0.00066242478479277243], (0.8, 120.0): [-0.48070356185330182, -0.017621686995755746, 0.0084009638803223801, -0.00069300383808949318], (0.8, inf): [-0.47926687718713606, -0.016476575352367202, 0.0086097059646591811, -0.00072160843492730911], (0.85, 2.0): [-0.53366806986381743, -0.098288178252723263, 0.026002333446289064, -0.0019567144268844896], (0.85, 3.0): [-0.48995919239619989, -0.077312722648418056, 0.019368984865418108, -0.0013449670192265796], (0.85, 4.0): [-0.46956079162382858, -0.063818518513946695, 0.015581608910696544, -0.0010264315084377606], (0.85, 5.0): [-0.45790853796153624, -0.054680511194530226, 0.013229852432203093, -0.00084248430847535898], (0.85, 6.0): [-0.4505070841695738, -0.048050936682873302, 0.011636407582714191, -0.00072491480033529815], (0.85, 7.0): [-0.44548337477336181, -0.042996612516383016, 0.010493052959891263, -0.00064528784792153239], (0.85, 8.0): [-0.44186624932664148, -0.039040005821657585, 0.0096479530794160544, -0.00058990874360967567], (0.85, 9.0): [-0.43914118689812259, -0.035875693030752713, 0.0090088804130628187, -0.00055071480339399694], (0.85, 10.0): [-0.43701255390953769, -0.033300997407157376, 0.0085172159355344848, -0.00052272770799695464], (0.85, 11.0): [-0.43530109064899053, -0.031174742038490313, 0.0081335619868386066, -0.00050268353809787927], (0.85, 12.0): [-0.43389220376610071, -0.02939618314990838, 0.007830626267772851, -0.00048836431712678222], (0.85, 13.0): [-0.43271026958463166, -0.027890759135246888, 0.0075886916668632936, -0.00047819339710596971], (0.85, 14.0): [-0.43170230265007209, -0.026604156062396189, 0.0073939099688705547, -0.00047109996854335419], (0.85, 15.0): [-0.43083160459377423, -0.025494228911600785, 0.0072358738657550868, -0.00046630677052262481], (0.85, 16.0): [-0.4300699280587239, -0.024529612608808794, 0.0071069227026219683, -0.00046323869860941791], (0.85, 17.0): [-0.42939734931902857, -0.023685025616054269, 0.0070011541609695891, -0.00046147954942994158], (0.85, 18.0): [-0.42879829041505324, -0.022940655682782165, 0.006914006369119409, -0.00046070877994711774], (0.85, 19.0): [-0.42826119448419875, -0.022280181781634649, 0.0068417746905826433, -0.00046066841214091982], (0.85, 20.0): [-0.42777654887094479, -0.021690909076747832, 0.0067817408643717969, -0.00046118620289068032], (0.85, 24.0): [-0.42622450033640852, -0.019869646711890065, 0.0066276799593494029, -0.00046668820637553747], (0.85, 30.0): [-0.42463810443233418, -0.018130114737381745, 0.0065344613060499164, -0.00047835583417510423], (0.85, 40.0): [-0.42299917804589382, -0.016498222901308417, 0.0065120558343578407, -0.00049656043685325469], (0.85, 60.0): [-0.42129387265810464, -0.014992121475265813, 0.0065657795990087635, -0.00052069705640687698], (0.85, 120.0): [-0.41951580476366368, -0.013615722489371183, 0.0066923911275726814, -0.00054846911649167492], (0.85, inf): [-0.41768751825428968, -0.012327525092266726, 0.0068664920569562592, -0.00057403720261753539], (0.9, 1.0): [-0.65851063279096722, -0.126716242078905, 0.036318801917603061, -0.002901283222928193], (0.9, 2.0): [-0.50391945369829139, -0.096996108021146235, 0.024726437623473398, -0.0017901399938303017], (0.9, 3.0): [-0.44799791843058734, -0.077180370333307199, 0.018584042055594469, -0.0012647038118363408], (0.9, 4.0): [-0.42164091756145167, -0.063427071006287514, 0.014732203755741392, -0.00094904174117957688], (0.9, 5.0): [-0.40686856251221754, -0.053361940054842398, 0.012041802076025801, -0.00072960198292410612], (0.9, 6.0): [-0.39669926026535285, -0.046951517438004242, 0.010546647213094956, -0.00062621198002366064], (0.9, 7.0): [-0.39006553675807426, -0.04169480606532109, 0.0093687546601737195, -0.00054648695713273862], (0.9, 8.0): [-0.38570205067061908, -0.037083910859179794, 0.0083233218526375836, -0.00047177586974035451], (0.9, 9.0): [-0.38190737267892938, -0.034004585655388865, 0.0077531991574119183, -0.00044306547308527872], (0.9, 10.0): [-0.37893272918125737, -0.031394677600916979, 0.0072596802503533536, -0.0004160518834299966], (0.9, 11.0): [-0.37692512492705132, -0.028780793403136471, 0.0066937909049060379, -0.00037420010136784526], (0.9, 12.0): [-0.37506345200129187, -0.026956483290567372, 0.0064147730707776523, -0.00036595383207062906], (0.9, 13.0): [-0.37339516122383209, -0.02543949524844704, 0.0061760656530197187, -0.00035678737379179527], (0.9, 14.0): [-0.37216979891087842, -0.02396347606956644, 0.0059263234465969641, -0.0003439784452550796], (0.9, 15.0): [-0.371209456600122, -0.022696132732654414, 0.0057521677184623147, -0.00033961108561770848], (0.9, 16.0): [-0.36958924377983338, -0.022227885445863002, 0.0057691706799383926, -0.00035042762538099682], (0.9, 17.0): [-0.36884224719083203, -0.021146977888668726, 0.0055957928269732716, -0.00034283810412697531], (0.9, 18.0): [-0.36803087186793326, -0.020337731477576542, 0.0054655378095212759, -0.00033452966946535248], (0.9, 19.0): [-0.3676700404163355, -0.019370115848857467, 0.0053249296207149655, -0.00032975528909580403], (0.9, 20.0): [-0.36642276267188811, -0.019344251412284838, 0.0054454968582897528, -0.00034868111677540948], (0.9, 24.0): [-0.36450650753755193, -0.017284255499990679, 0.0052337500059176749, -0.00034898202845747288], (0.9, 30.0): [-0.36251868940168608, -0.015358560437631397, 0.0050914299956134786, -0.00035574528891633978], (0.9, 40.0): [-0.36008886676510943, -0.014016835682905486, 0.0051930835959111514, -0.00038798316011984165], (0.9, 60.0): [-0.35825590690268061, -0.011991568926537646, 0.0050632208542414191, -0.00039090198974493085], (0.9, 120.0): [-0.35543612237284411, -0.011074403997811812, 0.0053504570752765162, -0.00043647137428074178], (0.9, inf): [-0.35311806343057167, -0.0096254020092145353, 0.0054548591208177181, -0.00045343916634968493], (0.95, 1.0): [-0.65330318136020071, -0.12638310760474375, 0.035987535130769424, -0.0028562665467665315], (0.95, 2.0): [-0.47225160417826934, -0.10182570362271424, 0.025846563499059158, -0.0019096769058043243], (0.95, 3.0): [-0.4056635555586528, -0.077067172693350297, 0.017789909647225533, -0.001182961668735774], (0.95, 4.0): [-0.37041675177340955, -0.063815687118939465, 0.014115210247737845, -0.00089996098435117598], (0.95, 5.0): [-0.35152398291152309, -0.052156502640669317, 0.010753738086401853, -0.0005986841939451575], (0.95, 6.0): [-0.33806730015201264, -0.045668399809578597, 0.0093168898952878162, -0.00051369719615782102], (0.95, 7.0): [-0.32924041072104465, -0.040019601775490091, 0.0080051199552865163, -0.00042054536135868043], (0.95, 8.0): [-0.32289030266989077, -0.035575345931670443, 0.0070509089344694669, -0.00035980773304803576], (0.95, 9.0): [-0.31767304201477375, -0.032464945930165703, 0.0064755950437272143, -0.0003316676253661824], (0.95, 10.0): [-0.31424318064708656, -0.029133461621153, 0.0057437449431074795, -0.00027894252261209191], (0.95, 11.0): [-0.31113589620384974, -0.02685115250591049, 0.0053517905282942889, -0.00026155954116874666], (0.95, 12.0): [-0.30848983612414582, -0.025043238019239168, 0.0050661675913488829, -0.00025017202909614005], (0.95, 13.0): [-0.3059212907410393, -0.023863874699213077, 0.0049618051135807322, -0.00025665425781125703], (0.95, 14.0): [-0.30449676902720035, -0.021983976741572344, 0.0045740513735751968, -0.00022881166323945914], (0.95, 15.0): [-0.30264908294481396, -0.02104880307520084, 0.0044866571614804382, -0.00023187587597844057], (0.95, 16.0): [-0.30118294463097917, -0.020160231061926728, 0.0044170780759056859, -0.00023733502359045826], (0.95, 17.0): [-0.30020013353427744, -0.018959271614471574, 0.0041925333038202285, -0.00022274025630789767], (0.95, 18.0): [-0.29857886556874402, -0.018664437456802001, 0.0042557787632833697, -0.00023758868868853716], (0.95, 19.0): [-0.29796289236978263, -0.017632218552317589, 0.0040792779937959866, -0.00022753271474613109], (0.95, 20.0): [-0.29681506554838077, -0.017302563243037392, 0.0041188426221428964, -0.00023913038468772782], (0.95, 24.0): [-0.29403146911167666, -0.015332330986025032, 0.0039292170319163728, -0.00024003445648641732], (0.95, 30.0): [-0.29080775563775879, -0.013844059210779323, 0.0039279165616059892, -0.00026085104496801666], (0.95, 40.0): [-0.28821583032805109, -0.011894686715666892, 0.0038202623278839982, -0.00026933325102031252], (0.95, 60.0): [-0.28525636737751447, -0.010235910558409797, 0.0038147029777580001, -0.00028598362144178959], (0.95, 120.0): [-0.28241065885026539, -0.0086103836327305026, 0.0038450612886908714, -0.00030206053671559411], (0.95, inf): [-0.27885570064169296, -0.0078122455524849222, 0.0041798538053623453, -0.0003469494881774609], (0.975, 1.0): [-0.65203598304297983, -0.12608944279227957, 0.035710038757117347, -0.0028116024425349053], (0.975, 2.0): [-0.46371891130382281, -0.096954458319996509, 0.023958312519912289, -0.0017124565391080503], (0.975, 3.0): [-0.38265282195259875, -0.076782539231612282, 0.017405078796142955, -0.0011610853687902553], (0.975, 4.0): [-0.34051193158878401, -0.063652342734671602, 0.013528310336964293, -0.00083644708934990761], (0.975, 5.0): [-0.31777655705536484, -0.051694686914334619, 0.010115807205265859, -0.00054517465344192009], (0.975, 6.0): [-0.30177149019958716, -0.044806697631189059, 0.008483551848413786, -0.00042827853925009264], (0.975, 7.0): [-0.29046972313293562, -0.039732822689098744, 0.007435356037378946, -0.00037562928283350671], (0.975, 8.0): [-0.28309484007368141, -0.034764904940713388, 0.0062932513694928518, -0.00029339243611357956], (0.975, 9.0): [-0.27711707948119785, -0.031210465194810709, 0.0055576244284178435, -0.00024663798208895803], (0.975, 10.0): [-0.27249203448553611, -0.028259756468251584, 0.00499112012528406, -0.00021535380417035389], (0.975, 11.0): [-0.26848515860011007, -0.026146703336893323, 0.0046557767110634073, -0.00020400628148271448], (0.975, 12.0): [-0.26499921540008192, -0.024522931106167097, 0.0044259624958665278, -0.00019855685376441687], (0.975, 13.0): [-0.2625023751891592, -0.022785875653297854, 0.004150277321193792, -0.00018801223218078264], (0.975, 14.0): [-0.26038552414321758, -0.021303509859738341, 0.0039195608280464681, -0.00017826200169385824], (0.975, 15.0): [-0.25801244886414665, -0.020505508012402567, 0.0038754868932712929, -0.00018588907991739744], (0.975, 16.0): [-0.25685316062360508, -0.018888418269740373, 0.0035453092842317293, -0.00016235770674204116], (0.975, 17.0): [-0.25501132271353549, -0.018362951972357794, 0.0035653933105288631, -0.00017470353354992729], (0.975, 18.0): [-0.25325045404452656, -0.017993537285026156, 0.0036035867405376691, -0.00018635492166426884], (0.975, 19.0): [-0.25236899494677928, -0.016948921372207198, 0.0034138931781330802, -0.00017462253414687881], (0.975, 20.0): [-0.25134498025027691, -0.016249564498874988, 0.0033197284005334333, -0.00017098091103245596], (0.975, 24.0): [-0.24768690797476625, -0.014668160763513996, 0.0032850791186852558, -0.00019013480716844995], (0.975, 30.0): [-0.24420834707522676, -0.012911171716272752, 0.0031977676700968051, -0.00020114907914487053], (0.975, 40.0): [-0.24105725356215926, -0.010836526056169627, 0.0030231303550754159, -0.00020128696343148667], (0.975, 60.0): [-0.23732082703955223, -0.0095442727157385391, 0.0031432904473555259, -0.00023062224109383941], (0.975, 120.0): [-0.23358581879594578, -0.0081281259918709343, 0.0031877298679120094, -0.00024496230446851501], (0.975, inf): [-0.23004105093119268, -0.0067112585174133573, 0.0032760251638919435, -0.00026244001319462992], (0.99, 1.0): [-0.65154119422706203, -0.1266603927572312, 0.03607480609672048, -0.0028668112687608113], (0.99, 2.0): [-0.45463403324378804, -0.098701236234527367, 0.024412715761684689, -0.0017613772919362193], (0.99, 3.0): [-0.36402060051035778, -0.079244959193729148, 0.017838124021360584, -0.00119080116484847], (0.99, 4.0): [-0.31903506063953818, -0.061060740682445241, 0.012093154962939612, -0.00067268347188443093], (0.99, 5.0): [-0.28917014580689182, -0.052940780099313689, 0.010231009146279354, -0.00057178339184615239], (0.99, 6.0): [-0.27283240161179012, -0.042505435573209085, 0.0072753401118264534, -0.00031314034710725922], (0.99, 7.0): [-0.25773968720546719, -0.039384214480463406, 0.0069120882597286867, -0.00032994068754356204], (0.99, 8.0): [-0.24913629282433833, -0.033831567178432859, 0.0055516244725724185, -0.00022570786249671376], (0.99, 9.0): [-0.24252380896373404, -0.029488280751457097, 0.0045215453527922998, -0.00014424552929022646], (0.99, 10.0): [-0.23654349556639986, -0.02705600214566789, 0.0041627255469343632, -0.00013804427029504753], (0.99, 11.0): [-0.23187404969432468, -0.024803662094970855, 0.0037885852786822475, -0.00012334999287725012], (0.99, 12.0): [-0.22749929386320905, -0.023655085290534145, 0.0037845051889055896, -0.00014785715789924055], (0.99, 13.0): [-0.22458989143485605, -0.021688394892771506, 0.0034075294601425251, -0.00012436961982044268], (0.99, 14.0): [-0.22197623872225777, -0.020188830700102918, 0.0031648685865587473, -0.00011320740119998819], (0.99, 15.0): [-0.2193924323730066, -0.019327469111698265, 0.0031295453754886576, -0.00012373072900083014], (0.99, 16.0): [-0.21739436875855705, -0.018215854969324128, 0.0029638341057222645, -0.00011714667871412003], (0.99, 17.0): [-0.21548926805467686, -0.017447822179412719, 0.0028994805120482812, -0.00012001887015183794], (0.99, 18.0): [-0.21365014687077843, -0.01688869353338961, 0.0028778031289216546, -0.00012591199104792711], (0.99, 19.0): [-0.21236653761262406, -0.016057151563612645, 0.0027571468998022017, -0.00012049196593780046], (0.99, 20.0): [-0.21092693178421842, -0.015641706950956638, 0.0027765989877361293, -0.00013084915163086915], (0.99, 24.0): [-0.20681960327410207, -0.013804298040271909, 0.0026308276736585674, -0.0001355061502101814], (0.99, 30.0): [-0.20271691131071576, -0.01206095288359876, 0.0025426138004198909, -0.00014589047959047533], (0.99, 40.0): [-0.19833098054449289, -0.010714533963740719, 0.0025985992420317597, -0.0001688279944262007], (0.99, 60.0): [-0.19406768821236584, -0.0093297106482013985, 0.0026521518387539584, -0.00018884874193665104], (0.99, 120.0): [-0.19010213174677365, -0.0075958207221300924, 0.0025660823297025633, -0.00018906475172834352], (0.99, inf): [-0.18602070255787137, -0.0062121155165363188, 0.0026328293420766593, -0.00020453366529867131], (0.995, 1.0): [-0.65135583544951825, -0.1266868999507193, 0.036067522182457165, -0.0028654516958844922], (0.995, 2.0): [-0.45229774013072793, -0.09869462954369547, 0.024381858599368908, -0.0017594734553033394], (0.995, 3.0): [-0.35935765236429706, -0.076650408326671915, 0.016823026893528978, -0.0010835134496404637], (0.995, 4.0): [-0.30704474720931169, -0.063093047731613019, 0.012771683306774929, -0.00075852491621809955], (0.995, 5.0): [-0.27582551740863454, -0.052533353137885791, 0.0097776009845174372, -0.00051338031756399129], (0.995, 6.0): [-0.25657971464398704, -0.043424914996692286, 0.0074324147435969991, -0.00034105188850494067], (0.995, 7.0): [-0.24090407819707738, -0.039591604712200287, 0.0068848429451020387, -0.00034737131709273414], (0.995, 8.0): [-0.23089540800827862, -0.034353305816361958, 0.0056009527629820111, -0.00024389336976992433], (0.995, 9.0): [-0.22322694848310584, -0.030294770709722547, 0.0046751239747245543, -0.00017437479314218922], (0.995, 10.0): [-0.21722684126671632, -0.026993563560163809, 0.0039811592710905491, -0.00013135281785826703], (0.995, 11.0): [-0.21171635822852911, -0.025156193618212551, 0.0037507759652964205, -0.00012959836685175671], (0.995, 12.0): [-0.20745332165849167, -0.023318819535607219, 0.0034935020002058903, -0.00012642826898405916], (0.995, 13.0): [-0.20426054591612508, -0.021189796175249527, 0.003031472176128759, -9.0497733877531618e-05], (0.995, 14.0): [-0.20113536905578902, -0.020011536696623061, 0.0029215880889956729, -9.571527213951222e-05], (0.995, 15.0): [-0.19855601561006403, -0.018808533734002542, 0.0027608859956002344, -9.2472995256929217e-05], (0.995, 16.0): [-0.19619157579534008, -0.017970461530551096, 0.0027113719105000371, -9.9864874982890861e-05], (0.995, 17.0): [-0.19428015140726104, -0.017009762497670704, 0.0025833389598201345, -9.6137545738061124e-05], (0.995, 18.0): [-0.19243180236773033, -0.01631617252107519, 0.0025227443561618621, -9.8067580523432881e-05], (0.995, 19.0): [-0.19061294393069844, -0.01586226613672222, 0.0025207005902641781, -0.00010466151274918466], (0.995, 20.0): [-0.18946302696580328, -0.014975796567260896, 0.0023700506576419867, -9.5507779057884629e-05], (0.995, 24.0): [-0.18444251428695257, -0.013770955893918012, 0.0024579445553339903, -0.00012688402863358003], (0.995, 30.0): [-0.18009742499570078, -0.011831341846559026, 0.0022801125189390046, -0.00012536249967254906], (0.995, 40.0): [-0.17562721880943261, -0.010157142650455463, 0.0022121943861923474, -0.000134542652873434], (0.995, 60.0): [-0.17084630673594547, -0.0090224965852754805, 0.0023435529965815565, -0.00016240306777440115], (0.995, 120.0): [-0.16648414081054147, -0.0074792163241677225, 0.0023284585524533607, -0.00017116464012147041], (0.995, inf): [-0.16213921875452461, -0.0058985998630496144, 0.0022605819363689093, -0.00016896211491119114], (0.999, 1.0): [-0.65233994072089363, -0.12579427445444219, 0.035830577995679271, -0.0028470555202945564], (0.999, 2.0): [-0.45050164311326341, -0.098294804380698292, 0.024134463919493736, -0.0017269603956852841], (0.999, 3.0): [-0.35161741499307819, -0.076801152272374273, 0.016695693063138672, -0.0010661121974071864], (0.999, 4.0): [-0.29398448788574133, -0.06277319725219685, 0.012454220010543127, -0.00072644165723402445], (0.999, 5.0): [-0.25725364564365477, -0.053463787584337355, 0.0099664236557431545, -0.00054866039388980659], (0.999, 6.0): [-0.23674225795168574, -0.040973155890031254, 0.0062599481191736696, -0.00021565734226586692], (0.999, 7.0): [-0.21840108878983297, -0.037037020271877719, 0.0055908063671900703, -0.00020238790479809623], (0.999, 8.0): [-0.2057964743918449, -0.032500885103194356, 0.0046441644585661756, -0.00014769592268680274], (0.999, 9.0): [-0.19604592954882674, -0.029166922919677936, 0.0040644333111949814, -0.00012854052861297006], (0.999, 10.0): [-0.18857328935948367, -0.026316705703161091, 0.0035897350868809275, -0.00011572282691335702], (0.999, 11.0): [-0.18207431428535406, -0.024201081944369412, 0.0031647372098056077, -8.1145935982296439e-05], (0.999, 12.0): [-0.17796358148991101, -0.021054306118620879, 0.0023968085939602055, -1.5907156771296993e-05], (0.999, 13.0): [-0.17371965962745489, -0.019577162950177709, 0.0022391783473999739, -2.0613023472812558e-05], (0.999, 14.0): [-0.16905298116759873, -0.01967115985443986, 0.0026495208325889269, -9.1074275220634073e-05], (0.999, 15.0): [-0.16635662558214312, -0.017903767183469876, 0.0022301322677100496, -5.1956773935885426e-05], (0.999, 16.0): [-0.16388776549525449, -0.016671918839902419, 0.0020365289602744382, -4.3592447599724942e-05], (0.999, 17.0): [-0.16131934177990759, -0.015998918405126326, 0.0019990454743285904, -4.8176277491327653e-05], (0.999, 18.0): [-0.15880633110376571, -0.015830715141055916, 0.0021688405343832091, -8.061825248932771e-05], (0.999, 19.0): [-0.15644841913314136, -0.015729364721105681, 0.0022981443610378136, -0.00010093672643417343], (0.999, 20.0): [-0.15516596606222705, -0.014725095968258637, 0.0021117117014292155, -8.8806880297328484e-05], (0.999, 24.0): [-0.14997437768645827, -0.012755323295476786, 0.0018871651510496939, -8.0896370662414938e-05], (0.999, 30.0): [-0.14459974882323703, -0.011247323832877647, 0.0018637400643826279, -9.6415323191606741e-05], (0.999, 40.0): [-0.13933285919392555, -0.0097151769692496587, 0.0018131251876208683, -0.00010452598991994023], (0.999, 60.0): [-0.13424555343804143, -0.0082163027951669444, 0.0017883427892173382, -0.00011415865110808405], (0.999, 120.0): [-0.12896119523040372, -0.0070426701112581112, 0.0018472364154226955, -0.00012862202979478294], (0.999, inf): [-0.12397213562666673, -0.0056901201604149998, 0.0018260689406957129, -0.00013263452567995485]} # p values that are defined in the A table p_keys = [.1,.5,.675,.75,.8,.85,.9,.95,.975,.99,.995,.999] # v values that are defined in the A table v_keys = range(2, 21) + [24, 30, 40, 60, 120, inf] def _isfloat(x): """ returns True if x is a float, returns False otherwise """ try: float(x) except: return False return True def _phi(p): """returns the pth quantile inverse norm""" return scipy.stats.norm.isf(p) def _ptransform(p): """function for p-value abcissa transformation""" return -1. / (1. + 1.5 * _phi((1. + p)/2.)) def _select_points(a, list_like): """ returns one above a, one below a, and the third closest point to a sorted in ascending order for quadratic interpolation. Assumes that points above and below a exist. """ foo = [x for x in list(list_like) if x-a <= 0] z = [min(foo, key=lambda x : abs(x-a))] foo = [x for x in list(list_like) if x-a > 0] z.append(min(foo, key=lambda x : abs(x-a))) foo = [x for x in list(list_like) if x not in z] z.append(min(foo, key=lambda x : abs(x-a))) return sorted(z) def _func(a, p, r, v): """ calculates f-hat for the coefficients in a, probability p, sample mean difference r, and degrees of freedom v. """ # eq. 2.3 f = a[0]*math.log(r-1.) + \ a[1]*math.log(r-1.)**2 + \ a[2]*math.log(r-1.)**3 + \ a[3]*math.log(r-1.)**4 # eq. 2.7 and 2.8 corrections if r == 3: f += -0.002 / (1. + 12. * _phi(p)**2) if v <= 4.364: f += 1./517. - 1./(312.*(v,1e38)[v==inf]) else: f += 1./(191.*(v,1e38)[v==inf]) return -f def _interpolate_p(p, r, v): """ interpolates p based on the values in the A table for the scalar value of r and the scalar value of v """ # interpolate p (v should be in table) # if .5 < p < .75 use linear interpolation in q # if p > .75 use quadratic interpolation in log(y + r/v) # by -1. / (1. + 1.5 * _phi((1. + p)/2.)) # find the 3 closest v values p0, p1, p2 = _select_points(p, p_keys) y0 = _func(A[(p0, v)], p0, r, v) + 1. y1 = _func(A[(p1, v)], p1, r, v) + 1. y2 = _func(A[(p2, v)], p2, r, v) + 1. y_log0 = math.log(y0 + float(r)/float(v)) y_log1 = math.log(y1 + float(r)/float(v)) y_log2 = math.log(y2 + float(r)/float(v)) # If p < .85 apply only the ordinate transformation # if p > .85 apply the ordinate and the abcissa transformation # In both cases apply quadratic interpolation if p > .85: p_t = _ptransform(p) p0_t = _ptransform(p0) p1_t = _ptransform(p1) p2_t = _ptransform(p2) # calculate derivatives for quadratic interpolation d2 = 2*((y_log2-y_log1)/(p2_t-p1_t) - \ (y_log1-y_log0)/(p1_t-p0_t))/(p2_t-p0_t) if (p2+p0)>=(p1+p1): d1 = (y_log2-y_log1)/(p2_t-p1_t) - 0.5*d2*(p2_t-p1_t) else: d1 = (y_log1-y_log0)/(p1_t-p0_t) + 0.5*d2*(p1_t-p0_t) d0 = y_log1 # interpolate value y_log = (d2/2.) * (p_t-p1_t)**2. + d1 * (p_t-p1_t) + d0 # transform back to y y = math.exp(y_log) - float(r)/float(v) elif p > .5: # calculate derivatives for quadratic interpolation d2 = 2*((y_log2-y_log1)/(p2-p1) - \ (y_log1-y_log0)/(p1-p0))/(p2-p0) if (p2+p0)>=(p1+p1): d1 = (y_log2-y_log1)/(p2-p1) - 0.5*d2*(p2-p1) else: d1 = (y_log1-y_log0)/(p1-p0) + 0.5*d2*(p1-p0) d0 = y_log1 # interpolate values y_log = (d2/2.) * (p-p1)**2. + d1 * (p-p1) + d0 # transform back to y y = math.exp(y_log) - float(r)/float(v) else: # linear interpolation in q and p q0 = math.sqrt(2) * -y0 * \ scipy.stats.t.isf((1.+p0)/2., (v,1e38)[v>1e38]) q1 = math.sqrt(2) * -y1 * \ scipy.stats.t.isf((1.+p1)/2., (v,1e38)[v>1e38]) d1 = (q1-q0)/(p1-p0) d0 = q0 # interpolate values q = d1 * (p-p0) + d0 # transform back to y y = -q / (math.sqrt(2) * \ scipy.stats.t.isf((1.+p)/2., (v,1e38)[v>1e38])) return y def _interpolate_v(p, r, v): """ interpolates v based on the values in the A table for the scalar value of r and th """ # interpolate v (p should be in table) # ordinate: y**2 # abcissa: 1./v # find the 3 closest v values v0, v1, v2 = _select_points(v, v_keys+([],[1])[p>=.90]) # y = f - 1. y0 = _func(A[(p,v0)], p, r, v0) + 1. y1 = _func(A[(p,v1)], p, r, v1) + 1. y2 = _func(A[(p,v2)], p, r, v2) + 1. # if v2 is inf set to a big number so interpolation # calculations will work if v2 > 1e38: v2 = 1e38 # calculate derivatives for quadratic interpolation d2 = 2.*((y2**2-y1**2)/(1./v2-1./v1) - \ (y0**2-y1**2)/(1./v0-1./v1)) / (1./v2-1./v0) if (1./v2 + 1./v0) >= (1./v1+1./v1): d1 = (y2**2-y1**2) / (1./v2-1./v1) - 0.5*d2*(1./v2-1./v1) else: d1 = (y1**2-y0**2) / (1./v1-1./v0) + 0.5*d2*(1./v1-1./v0) d0 = y1**2 # calculate y y = math.sqrt((d2/2.)*(1./v-1./v1)**2. + d1*(1./v-1./v1)+ d0) return y def _qsturng(p, r, v): # r is interpolated through the q to y here we only need to # account for when p and/or v are not found in the table. global A, p_keys, v_keys if p < .1 or p > .999: raise ValueError('p must be between .1 and .999') if p < .9: if v < 2: raise ValueError('v must be > 2 when p < .9') else: if v < 1: raise ValueError('v must be > 1 when p >= .9') if A.has_key((p,v)): f = _func(A[(p,v)], p, r, v) y = f + 1. elif p not in p_keys and v not in v_keys+([],[1])[p>=.90]: # apply bilinear (quadratic) interpolation # # p0,v2 + o + p1,v2 + p2,v2 # r2 # # 1 # - (p,v) # v x # # r1 # p0,v1 + o + p1,v1 + p2,v1 # # # p0,v0 + o r0 + p1,v0 + p2,v0 # # _ptransform(p) # # (p1 and v1 may be below or above (p,v). The algorithm # works in both cases. For diagramatic simplicity it is # shown as above) # # 1. at v0, v1, and v2 use quadratic interpolation # to find r0, r1, r2 # # 2. use r0, r1, r2 and quadratic interpolaiton # to find y and (p,v) # find the 3 closest v values v0, v1, v2 = _select_points(v, v_keys+([],[1])[p>=.90]) # find the 2 closest p values p0, p1, p2 = _select_points(p, p_keys) r0 = _interpolate_p(p, r, v0) r1 = _interpolate_p(p, r, v1) r2 = _interpolate_p(p, r, v2) # calculate derivatives for quadratic interpolation d2 = 2.*((r2**2-r1**2)/(1./v2-1./v1) - \ (r0**2-r1**2)/(1./v0-1./v1)) / (1./v2-1./v0) if (1./v2 + 1./v0) >= (1./v1+1./v1): d1 = (r2**2-r1**2) / (1./v2-1./v1) - 0.5*d2*(1./v2-1./v1) else: d1 = (r1**2-r0**2) / (1./v1-1./v0) + 0.5*d2*(1./v1-1./v0) d0 = r1**2 # calculate y y = math.sqrt((d2/2.)*(1./v-1./v1)**2. + d1*(1./v-1./v1)+ d0) elif v not in v_keys+([],[1])[p>=.90]: y = _interpolate_v(p, r, v) elif p not in p_keys: y = _interpolate_p(p, r, v) return math.sqrt(2) * -y * \ scipy.stats.t.isf((1.+p)/2., (v,1e38)[v>1e38]) # make a qsturng functinon that will accept list-like objects _vqsturng = np.vectorize(_qsturng) def qsturng(p, r, v): """ returns the q-value of the Studentized Range q-distribution as a function of the probability (p), number of sample means (r), and the degrees of freedom (v). """ if all(map(_isfloat, [p, r, v])): return _qsturng(p, r, v) return _vqsturng(p, r, v) import scipy.optimize def _psturng(q, r, v): opt_func = lambda p, r, v: abs(_qsturng(p, r, v) - q) return 1. - scipy.optimize.fminbound(opt_func, .1, .999, args=(r,v)) _vpsturng = np.vectorize(_psturng) def psturng(q, r, v): """ returns the probability for the Studentized q-distribution where the value q cooresponds to qsturng(1 - p, r, v) If .001 is returned the probability should be interpreted as, p <= .001. Likewise if .9 is returned the probability should be interpreted as, p >= .9. """ if all(map(_isfloat, [q, r, v])): return _psturng(q, r, v) return _vpsturng(q, r, v) ##p, r, v = .9, 10, 20 ##print ##print 'p and v interpolation' ##print '\t20\t22\t24' ##print '.75',qsturng(.75, r, 20),qsturng(.75, r, 22),qsturng(.75, r, 24) ##print '.85',qsturng(.85, r, 20),qsturng(.85, r, 22),qsturng(.85, r, 24) ##print '.90',qsturng(.90, r, 20),qsturng(.90, r, 22),qsturng(.90, r, 24) ##print ##print 'p and v interpolation' ##print '\t120\t500\tinf' ##print '.950',qsturng(.95, r, 120),qsturng(.95, r, 500),qsturng(.95, r, inf) ##print '.960',qsturng(.96, r, 120),qsturng(.96, r, 500),qsturng(.96, r, inf) ##print '.975',qsturng(.975, r, 120),qsturng(.975, r, 500),qsturng(.975, r, inf) ##print ##print 'p and v interpolation' ##print '\t40\t50\t60' ##print '.950',qsturng(.95, r, 40),qsturng(.95, r, 50),qsturng(.95, r, 60) ##print '.960',qsturng(.96, r, 40),qsturng(.96, r, 50),qsturng(.96, r, 60) ##print '.975',qsturng(.975, r, 40),qsturng(.975, r, 50),qsturng(.975, r, 60) ##print ##print 'p and v interpolation' ##print '\t20\t22\t24' ##print '.50',qsturng(.5, r, 20),qsturng(.5, r, 22),qsturng(.5, r, 24) ##print '.60',qsturng(.6, r, 20),qsturng(.6, r, 22),qsturng(.6, r, 24) ##print '.75',qsturng(.75, r, 20),qsturng(.75, r, 22),qsturng(.75, r, 24)
[ "dominic@nursix.org" ]
dominic@nursix.org
646357d0513992ba17b94e2a0ec1f976826b4817
f16a17a6faa673007f69b9892c4558e11958c55d
/cmddocs/articles.py
477cd909cf357279d062fd6df3261881a72a0240
[ "MIT" ]
permissive
TaurusOlson/cmddocs
6b707110774d9a7f7449e08c32846e8b263d4b48
7c84a7ba32b21c7a3d15304c4590f6eab56c572a
refs/heads/master
2021-01-20T22:50:51.512550
2015-01-08T11:33:26
2015-01-08T11:33:26
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,952
py
import os import re import tempfile import subprocess # Function definitions def list_articles(dir): "lists all articles in current dir and below" subprocess.call(["tree", dir]) def list_directories(dir): "lists all directories in current dir and below" subprocess.call(["tree", "-d", dir]) def change_directory(dir,datadir): "changes directory" d = os.path.join(os.getcwd(),dir) # dont cd out of datadir if not datadir in d: d = datadir # if empty, switch to datadir if not dir: d = datadir # switch to dir try: os.chdir(d) return d except OSError: print("Directory %s not found" % dir) def edit_article(article, directory, editor, repo, default_commit_msg): """edit an article within your docs""" # set paths a = os.path.join(directory, article) d = os.path.dirname(a) # create dir(s) if not os.path.isdir(d): os.makedirs(d) # start editor try: subprocess.call([editor, a]) except OSError: print "'%s' No such file or directory" % editor # commit into git try: repo.git.add(a) if repo.is_dirty(): msg = raw_input("Commit message: ") if not msg: msg = default_commit_msg repo.git.commit(m=msg) else: print "Nothing to commit" except: pass def view_article(article,dir,pager): "view an article within your docs" a = os.path.join(dir,article) # read original file try: article = open(a, "r") except IOError: print "Error: Could not find %s" % article return content = article.read() article.close() # create tmp file and convert markdown to ansi with tempfile.NamedTemporaryFile(delete=False) as tmp: h = re.compile('^#{3,5}\s*(.*)\ *$',re.MULTILINE) content = h.sub('\033[1m\033[37m\\1\033[0m', content) h = re.compile('^#{1,2}\s*(.*)\ *$',re.MULTILINE) content = h.sub('\033[4m\033[1m\033[37m\\1\033[0m', content) h = re.compile('^\ {4}(.*)',re.MULTILINE) content = h.sub('\033[92m\\1\033[0m', content) h = re.compile('~~~\s*([^~]*)~~~[^\n]*\n',re.DOTALL) content = h.sub('\033[92m\\1\033[0m', content) tmp.write(content) # start pager and cleanup tmp file afterwards # -fr is needed for showing binary+ansi colored files to # be properly displayed try: subprocess.call([pager, "-fr", tmp.name]) except OSError: print "'%s' No such file or directory" % pager try: os.remove(tmp.name) except OSError: print "Error: Could not remove %s" % tmp.name def delete_article(article,dir,repo): "delete an article" a = os.path.join(dir,article) try: repo.git.rm(a) repo.git.commit(m="%s deleted" % article) print("%s deleted" % article) except: if os.path.isdir(a): try: os.rmdir(a) print("Removed directory %s which was not under version control" % a) except OSError: print("Could not remove %s - its maybe not empty" % a) else: try: os.remove(a) print("Removed file %s which was not under version control" % a) except OSError: print("File %s could not be removed" % a) return def move_article(dir,args,repo): "move an article from source to destination" args = args.split() if len(args)!=2: print "Invalid usage\nUse: mv source dest" return a = os.path.join(dir,args[0]) e = os.path.join(dir,args[1]) d = os.path.dirname(e) # create dir(s) if not os.path.isdir(d): os.makedirs(d) # move file in git and commit repo.git.mv(a,e) repo.git.commit(m="Moved %s to %s" % (a,e)) print("Moved %s to %s" % (a,e)) def search_article(keyword, directory, datadir, exclude): """ search for a keyword in every article within your current directory and below. Much like recursive grep. """ c = 0 r = re.compile(keyword) for dirpath, dirs, files in os.walk(directory): dirs[:] = [d for d in dirs if d not in exclude] for fname in files: path = os.path.join(dirpath, fname) f = open(path, "rt") for i, line in enumerate(f): if r.search(line): c = c + 1 print "* \033[92m%s\033[39m: %s" % (os.path.relpath(path, datadir), line.rstrip('\n')) return "Results: %s" % c def show_log(args,repo): """ Show latest git logs with specified number of entries and maybe for a specific file. """ args = args.split() format="format:%C(blue)%h %Cgreen%C(bold)%ad %Creset%s" dateformat="short" if len(args) >= 1: if os.path.isfile(os.path.join(os.getcwd(), args[0])): file = args[0] try: count = args[1] print "Last %s commits for %s" % (count, file) print repo.git.log(file, pretty=format, n=count, date=dateformat) except IndexError: count = 10 print "Last %s commits for %s" % (count, file) print repo.git.log(file, pretty=format, n=count, date=dateformat) else: count = args[0] try: file = args[1] print "Last %s commits for %s" % (count, file) print repo.git.log(file, pretty=format, n=count, date=dateformat) except IndexError: print "Last %s commits" % count print repo.git.log(pretty=format, n=count, date=dateformat) elif len(args) == 0: count = 10 print "Last %s commits" % count print repo.git.log(pretty=format, n=count,date=dateformat)
[ "flo@noqqe.de" ]
flo@noqqe.de
64cddf5250ac60f94ef5c62aedfa3eb120d3e5f8
8ca70628ca811e08fb77b8e251fc8e5049486a65
/airbyte-integrations/bases/base-python/base_python/cdk/streams/exceptions.py
6727216dd5dd50496241a0890070cb87439e8f82
[ "MIT" ]
permissive
Tana8M/airbyte
a19544d2f7997ec7551793f7077d3e02bfe6ac84
49296ef657be272684c7259ed0d6be06e574dbe1
refs/heads/master
2023-04-15T15:04:22.849307
2021-04-23T23:12:55
2021-04-23T23:12:55
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,825
py
""" MIT License Copyright (c) 2020 Airbyte Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. """ from typing import Union import requests class BaseBackoffException(requests.exceptions.HTTPError): pass class UserDefinedBackoffException(BaseBackoffException): """ An exception that exposes how long it attempted to backoff """ def __init__(self, backoff: Union[int, float], request: requests.PreparedRequest, response: requests.Response): """ :param backoff: how long to backoff in seconds :param request: the request that triggered this backoff exception :param response: the response that triggered the backoff exception """ self.backoff = backoff super().__init__(request=request, response=response) class DefaultBackoffException(BaseBackoffException): pass
[ "noreply@github.com" ]
noreply@github.com
2a4fbb6eebb89b4e2f3e140179c0dd9d6b1d9665
b9ab57c699ab64c349098bba8ae3f954cd08b152
/quora-question-pairs/Model.py
0616eae7ed7f71ac72faf6cf5a75f06b53036b5d
[]
no_license
tyxr/my_tensorflow_code
c1e24eb3dc628b26ed7aeed92dcc3e0f8143fb0f
860acaee74a8ca6edcaf0ad3121399a94b79643e
refs/heads/master
2022-01-10T17:45:30.339149
2019-05-08T03:18:24
2019-05-08T03:18:24
106,514,006
1
0
null
null
null
null
UTF-8
Python
false
false
5,393
py
import tensorflow as tf from tensorflow.contrib.rnn import LSTMCell, DropoutWrapper from tensorflow.keras.layers import Embedding,TimeDistributed,Dense,Lambda from Process_data import print_shape import numpy as np class SimpleNetwork(): def __init__(self,num_words,embedding_dim,sequence_length,n_classes,L2,hidden_size,optimizer,learning_rate,clip_value): #para init self.num_words = num_words self.embedding_dim = embedding_dim self.sequence_length = sequence_length self.n_classes = n_classes self.L2 = L2 self.hidden_size = hidden_size self.optimizer = optimizer self.learning_rate = learning_rate self.clip_value = clip_value self._placeholder_init() # model operation self.logits = self._logits_op() self.loss = self._loss_op() self.acc = self._acc_op() self.train = self._training_op() tf.add_to_collection('train_mini', self.train) def _placeholder_init(self): self.q1 = tf.placeholder(tf.float32, [None, self.sequence_length], 'q1') self.q2 = tf.placeholder(tf.float32, [None, self.sequence_length], 'q2') self.y = tf.placeholder(tf.float32, None, 'y_true')#这里的1本身应该是类别,我们二分类就输出一个就好了。 #self.embed_matrix = tf.placeholder(tf.float32, [self.num_words+1, self.embedding_dim], 'embed_matrix') self.embed_matrix = np.load('./data/word_embedding_matrix.npy') self.dropout_keep_prob = tf.placeholder(tf.float32, name="dropout_keep_prob") def _logits_op(self): print(self.q1) print(self.embed_matrix) print(self.num_words+1) q1 = Embedding(self.num_words+1, self.embedding_dim,weights=[self.embed_matrix],input_length=self.sequence_length,trainable=False )(self.q1) q1 = TimeDistributed(Dense(self.embedding_dim, activation='relu'))(q1) q1 = Lambda(lambda x: tf.reduce_max(x, axis=1), output_shape=(self.embedding_dim,))(q1) q2 = Embedding(self.num_words+1, self.embedding_dim, weights=[self.embed_matrix],input_length=self.sequence_length, trainable=False )(self.q2) q2 = TimeDistributed(Dense(self.embedding_dim, activation='relu'))(q2) q2 = Lambda(lambda x: tf.reduce_max(x, axis=1), output_shape=(self.embedding_dim,))(q2) features = tf.concat([q1, q2], axis=1) logits = self._feedForwardBlock(features, self.hidden_size, self.n_classes, 'feed_forward') return logits def _loss_op(self): with tf.name_scope('cost'): losses = tf.nn.sigmoid_cross_entropy_with_logits(labels=self.y, logits=self.logits) loss = tf.reduce_mean(losses, name='loss_val') weights = [v for v in tf.trainable_variables() if ('w' in v.name) or ('kernel' in v.name)] l2_loss = tf.add_n([tf.nn.l2_loss(w) for w in weights]) * self.L2 loss += l2_loss return loss def _acc_op(self): with tf.name_scope('acc'): label_pred = self.logits label_true = self.y correct_pred = tf.equal(tf.round(label_pred), tf.round(label_true)) accuracy = tf.reduce_mean(tf.cast(correct_pred, tf.float32), name='Accuracy') return accuracy def _feedForwardBlock(self, inputs, hidden_dims, num_units, scope): """ :param inputs: tensor with shape (batch_size, 4 * 2 * hidden_size) :param scope: scope name :return: output: tensor with shape (batch_size, num_units) """ with tf.variable_scope(scope): initializer = tf.random_normal_initializer(0.0, 0.1) with tf.variable_scope('feed_foward_layer1'): inputs = tf.nn.dropout(inputs, self.dropout_keep_prob) outputs = tf.layers.dense(inputs, hidden_dims, tf.nn.relu, kernel_initializer = initializer) with tf.variable_scope('feed_foward_layer2'): outputs = tf.nn.dropout(outputs, self.dropout_keep_prob) results = tf.layers.dense(outputs, 1, tf.nn.sigmoid, kernel_initializer = initializer) return results def _training_op(self): with tf.name_scope('training'): if self.optimizer == 'adam': optimizer = tf.train.AdamOptimizer(self.learning_rate) elif self.optimizer == 'rmsprop': optimizer = tf.train.RMSPropOptimizer(self.learning_rate) elif self.optimizer == 'momentum': optimizer = tf.train.MomentumOptimizer(self.learning_rate, momentum=0.9) elif self.optimizer == 'sgd': optimizer = tf.train.GradientDescentOptimizer(self.learning_rate) elif self.optimizer == 'adadelta': optimizer = tf.train.AdadeltaOptimizer(self.learning_rate) elif self.optimizer == 'adagrad': optimizer = tf.train.AdagradOptimizer(self.learning_rate) else: ValueError('Unknown optimizer : {0}'.format(self.optimizer)) gradients, v = zip(*optimizer.compute_gradients(self.loss)) if self.clip_value is not None: gradients, _ = tf.clip_by_global_norm(gradients, self.clip_value) train_op = optimizer.apply_gradients(zip(gradients, v)) return train_op
[ "xingkai@icarbonx.com" ]
xingkai@icarbonx.com
7d718cf991da1edadb32ecc20b535482d1412c2f
9c6e055fc8c15d5a2e1dbec40cf2ab6534b49793
/nourisher/settings.py
264a24d78a7ca4f660a040e60b4b486ec63c91ab
[]
no_license
hnykda/nourisher
0d70bac6a19ed95656e7ef51a4bb18e8b4aae9d2
6049d5b47c9004b3badfee89bcf7c259c84c4442
refs/heads/master
2021-01-20T16:12:04.940861
2015-07-27T16:42:04
2015-07-27T16:42:04
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,541
py
import logging log = logging.getLogger(__name__) """ Created on Dec 22, 2014 @author: dan """ """ Here is possible to set customizables - global variables """ ### DATABASE ### DB_IP = "localhost" # IP where is MongoDB running DB_PORT = 5432 # port where is database running DB_NAME = "testdb" # Name of default database DB_COLLECTION = "feeds" # Name of collection in database ### MATERNALSITE ### # which selenium.webdriver and settings # should be used for scrapping. Possible values in maternaSite.Scraper DEFAULT_DRIVER = "phantomjs" ARTICLES_LIMIT = 25 ### VERBOSITY ### #VERBOSITY = 1 # Verbosity of std output (currently implemented 0, 1, 2) def get_setings(): """Print current settings""" log.debug(DB_COLLECTION + " " + str(DB_PORT) + " " + DB_IP + " " + DB_NAME + " " + DEFAULT_DRIVER) # class SETTER: # ### DATABASE ### # DB_IP = "localhost" # IP where is MongoDB running # DB_PORT = 5432 # port where is database running # DB_NAME = "testdb" # Name of default database # DB_COLLECTION = "feeds" # Name of collection in database # # # ### MATERNALSITE ### # # which selenium.webdriver and settings # # should be used for scrapping. Possible values in maternaSite.Scraper # DEFAULT_DRIVER = "phantomjs" # # ### VERBOSITY ### # VERBOSITY = 1 # Verbosity of std output (currently implemented 0, 1, 2) # # def set_db_collection( self, name ): # self.DB_COLLECTION = name # # def get_db_collection( self ): # return( self.DB_COLLECTION )
[ "kotrfa@gmail.com" ]
kotrfa@gmail.com
be9bfbbae588fe4bff415889569809c937fc371a
c6ac57175975de3353faee0c2f1f796e2cd16a16
/bot1.py
4ee830465388df215de35511828cc2758ed29857
[]
no_license
3rdmonth/env_pizzabot
9c37062567dbdd010b40d1eb19de91f96cdd9b9a
ebe8bdad23206514a5f7ac3c6a3fca3b302bde78
refs/heads/master
2022-12-14T14:23:47.930610
2018-09-29T18:22:03
2018-09-29T18:22:03
150,890,812
0
0
null
2021-06-01T22:42:00
2018-09-29T18:12:21
Python
UTF-8
Python
false
false
689
py
import requests url = "https://api.telegram.org/bot645781983:AAHslgjt-LRnVX57f-40hlQwQMBC-oYoQtc/" def get_updates_json(request): response = requests.get(request + 'getUpdates') return response.json() def last_update(data): results = data['result'] total_updates = len(results) - 1 return results[total_updates] def get_chat_id(update): chat_id = update['message']['chat']['id'] return chat_id def send_mess(chat, text): params = {'chat_id': chat, 'text': text} response = requests.post(url + 'sendMessage', data=params) return response chat_id = get_chat_id(last_update(get_updates_json(url))) send_mess(chat_id, 'Your message goes here')
[ "vas.viktorov@ya.ru" ]
vas.viktorov@ya.ru
ddc6ecb74c82d9731a0332ca8a4960eb0ace0285
b78b3cb27b84b79246dd24b88108a50d7e52edb6
/pprzrest/mSimProcMon.py
8c89289003c555db1ba0caef5ff130560f99494f
[]
no_license
savass/paparazzi_box_frontend
93012f2f0919785c41b5a14509c1a6103f1605ce
1b91659b890d16f897b63dc224836af29d0d2f06
refs/heads/master
2021-01-25T07:34:31.278819
2015-04-22T13:43:44
2015-04-22T13:43:44
34,392,545
0
0
null
null
null
null
UTF-8
Python
false
false
5,442
py
#Simple monitoring tool to watch agents import time from threading import Thread import fcntl import os from mProcessHolder import ProcessHolder class SimProcMon: ProcessDemonThread = None RunProcessDeamon = True ProcessList = [] ProcessOutput = "" def __init__(self, Logger, SocketIo=None): self.mLog = Logger self.SocketIo = SocketIo self.mLog.debug("New SimProcMon class created.") def inform_clients(self): self.mLog.debug("Informing clients..") self.SocketIo.emit('StatusMsg',{'data': self.ProcessOutput, 'link_sts': self.get_status("link"), 'server_sts': self.get_status("server"), 'app_server_sts': self.get_status("app_server"), 'link_arg': self.get_process_arg("link"), 'server_arg':self.get_process_arg("server"), 'app_server_arg': self.get_process_arg("app_server")},namespace='/PprzOnWeb') def get_status(self, ProcessName): for mProcess in self.ProcessList: if mProcess.ProcName == ProcessName: return mProcess.ProcessRunning return False def stop_process(self, ProcessName): for mProcess in self.ProcessList: if mProcess.ProcName == ProcessName: mProcess.RunProcess = False if not mProcess.ProcessAgent is None: mProcess.ProcessAgent.kill() mProcess.ProcessAgent = None self.mLog.info("%s agent terminated.", mProcess.ProcName) #remove from ProcessList self.ProcessList.remove(mProcess) self.ProcessOutput = self.ProcessOutput + mProcess.ProcName + " agent terminated."+ "<br />" self.inform_clients() return else: self.mLog.error("%s process agent is None.", mProcess.ProcName) def start_process(self, ProcessName, RunStr): #need to check if process already started. if self.get_status(ProcessName): self.mLog.error("%s is already started.", ProcessName) return nPH = ProcessHolder(ProcessName, RunStr, True) nPH.run() #append the process to process list self.ProcessList.append(nPH) self.mLog.info("Trying to start %s agent.", ProcessName) def process_deamon(self): self.mLog.info("Process monitor started.") while self.RunProcessDeamon: time.sleep(0.5) for mProcess in self.ProcessList: #first need to check if process is running or not. if mProcess.RunProcess and not mProcess.ProcessAgent is None: #print "mProcess ", mProcess.ProcessRunning fd = mProcess.ProcessAgent.stdout.fileno() fl = fcntl.fcntl(fd, fcntl.F_GETFL) fcntl.fcntl(fd, fcntl.F_SETFL, fl | os.O_NONBLOCK) try: AppOut = mProcess.ProcessAgent.stdout.read() if AppOut: self.ProcessOutput = self.ProcessOutput + "<br />".join(AppOut.split("\n")) self.inform_clients() except: pass if not mProcess.ProcessRunning: #give some time to process self.mLog.info("%s seems to be fresh started. Waiting..", mProcess.ProcName) time.sleep(1.5) if not mProcess.ProcessAgent.poll() is None: #process error <span class="myClass">test</span> self.ProcessOutput = self.ProcessOutput + '<span class="MonError">' + mProcess.ProcName + ' stopped.</span> <br />' if mProcess.RunProcess and mProcess.ProcessRunning: mProcess.run() self.ProcessOutput = self.ProcessOutput + '<span class="MonOutput">' + mProcess.ProcName + ' restarted. </span> <br />' self.mLog.info("%s.ProcessAgent.poll() is None > ProcessRunning.", mProcess.ProcName) self.inform_clients() else: self.mLog.info("%s.ProcessAgent.poll() is None > startup failed!", mProcess.ProcName) self.ProcessList.remove(mProcess) self.inform_clients() continue #seems like process started successfully if not mProcess.ProcessRunning: mProcess.ProcessRunning = True self.ProcessOutput = self.ProcessOutput + '<span class="MonOutput">' + mProcess.ProcName + ' agent started.. </span> <br />' self.mLog.info("%s agent started.", mProcess.ProcName) self.inform_clients() #print len(ProcessList) self.mLog.info("Process monitor stopped.") def get_process_arg(self,ProcessName): for mProcess in self.ProcessList: if mProcess.ProcName == ProcessName: return mProcess.ProcArg return None def run_process_mon_deamon(self): if self.ProcessDemonThread is None: self.ProcessDemonThread = Thread(target=self.process_deamon) self.ProcessDemonThread.daemon = True self.ProcessDemonThread.start() else: self.mLog.error("Process monitor deamon already stared!")
[ "sen.savas@gmail.com" ]
sen.savas@gmail.com
f67485b6750b8322f509252eb0862cc7b39aa952
38c4f683d6fe10d671834549fe80406b1de75ced
/api/users/migrations/0001_initial.py
830889b99ddfc74ce22d6d468ae8a4c0f8853664
[ "BSD-2-Clause" ]
permissive
charles-vdulac/django-rest-skeleton
1a3b05e349eaad82e0e99477aedc8dad4ff1698a
2e621c030046f317b01359b716dae2120e1f7525
refs/heads/master
2020-12-24T10:53:55.305379
2014-06-12T13:09:12
2014-06-12T13:09:12
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,961
py
# -*- coding: utf-8 -*- import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): needed_by = ( ("authtoken", "0001_initial"), ) def forwards(self, orm): # Adding model 'User' db.create_table(u'users_user', ( (u'id', self.gf('django.db.models.fields.AutoField')(primary_key=True)), ('password', self.gf('django.db.models.fields.CharField')(max_length=128)), ('last_login', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now)), ('is_superuser', self.gf('django.db.models.fields.BooleanField')(default=False)), ('uid', self.gf('django.db.models.fields.CharField')(max_length=36, blank=True)), ('email', self.gf('django.db.models.fields.EmailField')(unique=True, max_length=254)), ('first_name', self.gf('django.db.models.fields.CharField')(max_length=30, blank=True)), ('last_name', self.gf('django.db.models.fields.CharField')(max_length=30, blank=True)), ('is_staff', self.gf('django.db.models.fields.BooleanField')(default=False)), ('is_active', self.gf('django.db.models.fields.BooleanField')(default=True)), ('date_joined', self.gf('django.db.models.fields.DateTimeField')(default=datetime.datetime.now)), )) db.send_create_signal(u'users', ['User']) # Adding M2M table for field groups on 'User' m2m_table_name = db.shorten_name(u'users_user_groups') db.create_table(m2m_table_name, ( ('id', models.AutoField(verbose_name='ID', primary_key=True, auto_created=True)), ('user', models.ForeignKey(orm[u'users.user'], null=False)), ('group', models.ForeignKey(orm[u'auth.group'], null=False)) )) db.create_unique(m2m_table_name, ['user_id', 'group_id']) # Adding M2M table for field user_permissions on 'User' m2m_table_name = db.shorten_name(u'users_user_user_permissions') db.create_table(m2m_table_name, ( ('id', models.AutoField(verbose_name='ID', primary_key=True, auto_created=True)), ('user', models.ForeignKey(orm[u'users.user'], null=False)), ('permission', models.ForeignKey(orm[u'auth.permission'], null=False)) )) db.create_unique(m2m_table_name, ['user_id', 'permission_id']) def backwards(self, orm): # Deleting model 'User' db.delete_table(u'users_user') # Removing M2M table for field groups on 'User' db.delete_table(db.shorten_name(u'users_user_groups')) # Removing M2M table for field user_permissions on 'User' db.delete_table(db.shorten_name(u'users_user_user_permissions')) models = { u'auth.group': { 'Meta': {'object_name': 'Group'}, u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'unique': 'True', 'max_length': '80'}), 'permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}) }, u'auth.permission': { 'Meta': {'ordering': "(u'content_type__app_label', u'content_type__model', u'codename')", 'unique_together': "((u'content_type', u'codename'),)", 'object_name': 'Permission'}, 'codename': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'content_type': ('django.db.models.fields.related.ForeignKey', [], {'to': u"orm['contenttypes.ContentType']"}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '50'}) }, u'contenttypes.contenttype': { 'Meta': {'ordering': "('name',)", 'unique_together': "(('app_label', 'model'),)", 'object_name': 'ContentType', 'db_table': "'django_content_type'"}, 'app_label': ('django.db.models.fields.CharField', [], {'max_length': '100'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'model': ('django.db.models.fields.CharField', [], {'max_length': '100'}), 'name': ('django.db.models.fields.CharField', [], {'max_length': '100'}) }, u'users.user': { 'Meta': {'object_name': 'User'}, 'date_joined': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'email': ('django.db.models.fields.EmailField', [], {'unique': 'True', 'max_length': '254'}), 'first_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'groups': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['auth.Group']", 'symmetrical': 'False', 'blank': 'True'}), u'id': ('django.db.models.fields.AutoField', [], {'primary_key': 'True'}), 'is_active': ('django.db.models.fields.BooleanField', [], {'default': 'True'}), 'is_staff': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'is_superuser': ('django.db.models.fields.BooleanField', [], {'default': 'False'}), 'last_login': ('django.db.models.fields.DateTimeField', [], {'default': 'datetime.datetime.now'}), 'last_name': ('django.db.models.fields.CharField', [], {'max_length': '30', 'blank': 'True'}), 'password': ('django.db.models.fields.CharField', [], {'max_length': '128'}), 'uid': ('django.db.models.fields.CharField', [], {'max_length': '36', 'blank': 'True'}), 'user_permissions': ('django.db.models.fields.related.ManyToManyField', [], {'to': u"orm['auth.Permission']", 'symmetrical': 'False', 'blank': 'True'}) } } complete_apps = ['users']
[ "sebastibe@gmail.com" ]
sebastibe@gmail.com
284cf89f5485e9276d982fb582c9377e25489e30
920a5a8df0a6afbdcc1363fb74292074dec3b516
/tf2onnx/rewriter/random_uniform.py
23c450964d2db499b4dcc134d45f931bca7fd276
[ "MIT" ]
permissive
anttisaukko/tensorflow-onnx
a674a6ce0002912c42914a69e8333b379391af46
1341bdf476df6023b75bc6b3c6e4cda00cc58a29
refs/heads/master
2020-04-08T21:17:35.616038
2018-11-28T14:47:27
2018-11-28T14:47:27
159,737,860
0
0
MIT
2018-11-29T22:54:42
2018-11-29T22:54:41
null
UTF-8
Python
false
false
3,044
py
# Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT license. """ tf2onnx.rewrite - rewrite tensorflow subgraph to onnx random_uniform op """ from onnx import helper from tf2onnx.graph import Node from tf2onnx.graph_matcher import OpTypePattern, GraphMatcher from tf2onnx import utils from tf2onnx.utils import port_name # pylint: disable=missing-docstring def rewrite_random_uniform(g, ops): pattern = \ OpTypePattern('Add', name='output', inputs=[ OpTypePattern('Mul', inputs=[ OpTypePattern('RandomUniform', name='input1', inputs=["*"]), OpTypePattern('Sub', name='input2', inputs=["*", "*"]), ]), None ]) matcher = GraphMatcher(pattern) match_results = list(matcher.match_ops(ops)) for match in match_results: input2 = match.get_op('input2') output = match.get_op('output') ru_op = match.get_op('input1') # max is on input 0 tmax = input2.inputs[0].get_tensor_value()[0] tmin = input2.inputs[1].get_tensor_value()[0] new_node = create_onnx_random_uniform_op(g, tmax, tmin, ru_op, output) ops = g.replace_subgraph(ops, match, [], [output], [], [new_node]) return ops # rewriter function when fold_const is enabled def rewrite_random_uniform_fold_const(g, ops): pattern = \ OpTypePattern('Add', name='output', inputs=[ OpTypePattern('Mul', name='mul', inputs=[ OpTypePattern('RandomUniform', name='input1', inputs=["*"]), None, ]), None, ]) matcher = GraphMatcher(pattern) match_results = list(matcher.match_ops(ops)) for match in match_results: output = match.get_op('output') mul = match.get_op('mul') ru_op = match.get_op('input1') tmax_minus_tmin = mul.inputs[1].get_tensor_value()[0] tmin = output.inputs[1].get_tensor_value()[0] tmax = tmin + tmax_minus_tmin new_node = create_onnx_random_uniform_op(g, tmax, tmin, ru_op, output) ops = g.replace_subgraph(ops, match, [], [output], [], [new_node]) return ops def create_onnx_random_uniform_op(g, tmax, tmin, ru_op, output): dtype = output.dtype op_name = utils.make_name("RandomUniform") out_name = port_name(op_name) if ru_op.inputs[0].type == "Shape": shape_node = ru_op.inputs[0] new_node = Node(helper.make_node("RandomUniformLike", [shape_node.input[0]], [out_name], name=op_name, low=tmin, high=tmax, dtype=dtype), g) else: shape = g.get_shape(output.output[0]) new_node = Node(helper.make_node("RandomUniform", [], [out_name], name=op_name, low=tmin, high=tmax, dtype=dtype, shape=shape), g) return new_node
[ "pengwa@microsoft.com" ]
pengwa@microsoft.com
b02ec88d3c3a9fda595bfaded4054a2a7b2b905d
a1ff3c1e1f633f8c793c69214bf70ed6321891f2
/frequencyAnalyzer.py
b163360268f399de2fb79d30583ca65f3292d0bb
[]
no_license
yanone/frequencyAnalyzer
6f857a72a5e3a2b64b01a99816dd6ac33c3580be
0312f21a0662db3524382f80b88dea28d437cd65
refs/heads/master
2021-09-23T03:26:46.349196
2018-09-20T08:39:43
2018-09-20T08:39:43
115,872,586
4
0
null
null
null
null
UTF-8
Python
false
false
12,520
py
import plistlib import threading import pyaudio import numpy as np import wave import audioop import time, os import math import json import wx from ynlib.maths import Interpolate #from pysine import sine ############################################################### frequencies = [] interpolatedFrequencies = [] volumes = {} clipping = {} intermediateSteps = 3 minimumVolume = None maximumVolume = None averageVolume = 0 currentVolume = 0 peakVolume = 80 volumeScope = 120.0 ############################################################### # output out = pyaudio.PyAudio() _volume = 1.0 # range [0.0, 1.0] fs = 44100 # sampling rate, Hz, must be integer duration = 0.10 # in seconds, may be float # for paFloat32 sample values must be in range [-1.0, 1.0] # input CHUNK = 4096 FORMAT = pyaudio.paInt16 CHANNELS = 1 RATE = 44100 ############################################################### def CleanFloat(number, locale = 'en'): """\ Return number without decimal points if .0, otherwise with .x) """ try: if number % 1 == 0: return str(int(number)) else: return str(float(number)) except: pass ############################################################### outputStream = out.open(format=pyaudio.paFloat32, channels=CHANNELS, rate=RATE, output=True, frames_per_buffer=2048, ) class AppKitNSUserDefaults(object): def __init__(self, name = None): from AppKit import NSUserDefaults if name: self.defaults = NSUserDefaults.alloc().initWithSuiteName_(name) else: self.defaults = NSUserDefaults.standardUserDefaults() def get(self, key): if self.defaults.objectForKey_(key): return json.loads(self.defaults.objectForKey_(key)) def set(self, key, value): self.defaults.setObject_forKey_(json.dumps(value), key) def remove(self, key): self.defaults.removeObjectForKey_(key) def play(f, thread): # generate samples, note conversion to float32 array # samples = (np.sin(2*np.pi*np.arange(fs*duration*2.0)*f/fs)).astype(np.float32) samples = (np.sin(2*np.pi*np.arange(fs*duration)*f/fs)).astype(np.float32).tobytes() # print samples # def sine_wave(frequency=440.0, framerate=RATE, amplitude=0.5): # amplitude = max(min(amplitude, 1), 0) # return (float(amplitude) * math.sin(2.0*math.pi*float(frequency)*(float(i)/float(framerate))) for i in count(0)) # samples = (f, RATE) # print samples def sine_wave(frequency=440.0, framerate=44100, amplitude=0.5, duration = 1.0): period = int(framerate / frequency) if amplitude > 1.0: amplitude = 1.0 if amplitude < 0.0: amplitude = 0.0 lookup_table = [float(amplitude) * math.sin(2.0*math.pi*float(frequency)*(float(i%period)/float(framerate))) for i in range(period)] return lookup_table(lookup_table[i%period] for i in range(period)) #eq # samples = sine_wave(f, RATE, .5, .1) # print samples # ramp = int(len(samples) * .1) # # print samples[-ramp:] # for i in range(ramp): # samples[i] = Interpolate(0, samples[i], i/float(ramp)) # samples[-i-1] = Interpolate(0, samples[-i-1], i/float(ramp)) # print samples[-ramp:] # play. May repeat with different volume values (if done interactively) outputStream.write(samples) def volume(f, thread): # print f global averageVolume, clipping, currentVolume time.sleep(max(0, duration / 2.0 )) input = pyaudio.PyAudio() inputStream = input.open(format=FORMAT, channels=CHANNELS, rate=RATE, input=True, frames_per_buffer=CHUNK) _max = 0 values = [] for i in range(0, int(RATE / CHUNK * max(.1, duration * .2))): data = inputStream.read(CHUNK) rms = audioop.rms(data, 2) # here's where you calculate the volume values.append(rms) _max = max(_max, rms) # for i in range(10): #to it a few times just to see # data = np.fromstring(inputStream.read(CHUNK),dtype=np.int16) # data = data * np.hanning(len(data)) # smooth the FFT by windowing data # fft = abs(np.fft.fft(data).real) # fft = fft[:int(len(fft)/2)] # keep only first half # freq = np.fft.fftfreq(CHUNK,1.0/RATE) # freq = freq[:int(len(freq)/2)] # keep only first half # freqPeak = freq[np.where(fft==np.max(fft))[0][0]]+1 # print(f, "peak frequency: %d Hz"%freqPeak) # value = 0 value = sum(values) / float(len(values)) # print(value) # dB(A) value = 20 * math.log10(value) + 2.0 if value > peakVolume: clipping[f] = True else: clipping[f] = False currentVolume = value # print value # print f, _max volumes[f] = value averageVolume = sum(volumes.values())/ float(len(list(volumes.values()))) thread.frame._max = max(thread.frame._max, value) # thread.frame.Refresh() # print volumes # print min(volumes.values()), max(volumes.values()) inputStream.stop_stream() inputStream.close() input.terminate() class Record(threading.Thread): def __init__(self, frame): threading.Thread.__init__(self) self.frame = frame def run(self): while self.frame.alive: for f in interpolatedFrequencies: if self.frame.playing: self.frame.currentFrequency = f self.frame.Refresh() p = threading.Thread(target=play, args=(f, self,)) p.start() v = threading.Thread(target=volume, args=(f, self,)) v.start() time.sleep(duration) # for f in interpolatedFrequencies: # print '%s: %s' % (f, volumes[f]) time.sleep(duration) class Example(wx.Frame): def __init__(self, parent, title): super(Example, self).__init__(parent, title=title, size=(1000, 600)) self.preferences = AppKitNSUserDefaults('de.yanone.frequencyAnalyzer') self.alive = True self._max = 0 self.currentFrequency = None self.playing = False self.recorder = Record(self) self.recorder.start() self.deviceButton = wx.Button(self, -1, "Device") self.deviceButton.Bind(wx.EVT_BUTTON, self.OnDevice) self.playButton = wx.Button(self, -1, "Play") self.playButton.Bind(wx.EVT_BUTTON, self.OnPlay) self.stopButton = wx.Button(self, -1, "Stop") self.stopButton.Bind(wx.EVT_BUTTON, self.OnStop) self.Centre() self.Bind(wx.EVT_CLOSE, self.OnClose) self.Bind(wx.EVT_PAINT, self.OnPaint) dc = wx.ClientDC(self) dc.DrawLine(50, 60, 190, 60) if self.preferences.get('deviceFile'): self.openDeviceFile(self.preferences.get('deviceFile')) def OnPaint(self, event=None): dc = wx.PaintDC(self) dc.Clear() dc.SetPen(wx.Pen(wx.BLACK, 4)) size = dc.GetSize() dc.SetBackground(wx.Brush(wx.Colour(30,30,30))) dc.Clear() # dc.DrawRectangle(0, 0, size[0], size[1]) marginHorizontal = max(size[0] * .05, 100) marginTop = max(size[1] * .1, 100) marginBottom = max(size[1] * .2, 200) left = marginHorizontal right = size[0] - marginHorizontal - 100 top = marginTop bottom = size[1] - marginBottom height = max(1, (bottom - top)) width = max(1, (right - left)) colour = wx.Colour(223,219,0) colour = wx.Colour(223,219,0) activeColour = wx.Colour(229,53,45) fontSize = max(width / 80.0, 10) # dB(A) label font = wx.Font(fontSize, wx.FONTFAMILY_DEFAULT, wx.FONTSTYLE_NORMAL, wx.FONTWEIGHT_NORMAL) dc.SetTextForeground(colour) dc.SetFont(font) dc.DrawLabel('dB(A)', wx.Rect(left - 50, top + height / 2.0, 40, 20), wx.ALIGN_RIGHT) # Volume pen=wx.Pen(colour ,4) dc.SetPen(pen) y = bottom - height * currentVolume / volumeScope dc.DrawLine(right + 85, bottom, right + 85, y) pen=wx.Pen(activeColour ,4) dc.SetPen(pen) dc.DrawLine(right + 70, top, right + 100, top) dc.DrawLine(right + 100, top, right + 100, bottom) dc.DrawLine(right + 70, bottom, right + 100, bottom) y = bottom - height * peakVolume / volumeScope dc.DrawLine(right + 70, y, right + 100, y) if frequencies: _min = min(volumes.values()) _max = max(volumes.values()) # self._max = max(_max, self._max) if self._max: factor = float(height) / float(self._max) else: factor = 1 factor = height / volumeScope # print 'max', self._max, 'height', height, 'factor', factor # print factor # Peak y = bottom - peakVolume * factor pen=wx.Pen(activeColour ,4) dc.SetPen(pen) dc.DrawLine(left, y, right, y) for i, f in enumerate(interpolatedFrequencies): x = left + i * (right - left) / float(len(interpolatedFrequencies) - 1) # Average x2 = left + (i+1) * (right - left) / float(len(interpolatedFrequencies) - 1) if f <= 80: volumeAdjust = 12 elif f <= 100: volumeAdjust = 9 elif f <= 125: volumeAdjust = 6 elif f <= 160: volumeAdjust = 3 else: volumeAdjust = 0 y = bottom - (averageVolume + volumeAdjust) * factor pen=wx.Pen(wx.Colour(125,125,125) ,4) dc.SetPen(pen) dc.DrawLine(x, y, x2, y) # Grid if f in frequencies: if f in clipping and clipping[f] == True: pen=wx.Pen(colour ,4) else: pen=wx.Pen(activeColour ,4) dc.SetPen(pen) dc.DrawLine(x, top, x, bottom) # connecting lines pointPosition = (x, bottom - volumes[f] * factor) if i > 0: previousPointPosition = (left + (i-1) * (right - left) / float(len(interpolatedFrequencies) - 1), bottom - volumes[interpolatedFrequencies[i-1]] * factor) pen=wx.Pen(colour ,4) dc.SetPen(pen) dc.DrawLine(pointPosition[0], pointPosition[1], previousPointPosition[0], previousPointPosition[1]) if f in frequencies: font = wx.Font(fontSize, wx.FONTFAMILY_DEFAULT, wx.FONTSTYLE_NORMAL, wx.FONTWEIGHT_NORMAL) dc.SetTextForeground(colour) dc.SetFont(font) text = str(f) if f >= 1000: text = '%sk' % (f // 1000) if f % 1000: text += CleanFloat(f % 1000 / 100) dc.DrawLabel(text, wx.Rect(x - 20, bottom + max(20, size[0] / 75.0), 40, 20), wx.ALIGN_CENTER) if f in frequencies: dc.SetPen(wx.Pen(colour ,0)) dc.SetBrush(wx.Brush(colour)) pointSize = float(width) / float(len(frequencies) - 1) * .2 dc.DrawCircle(pointPosition[0], pointPosition[1], pointSize) self.deviceButton.SetPosition((marginHorizontal, size[1] - marginBottom + max(100, size[0] / 15.0))) self.playButton.SetPosition((marginHorizontal + 100, size[1] - marginBottom + max(100, size[0] / 15.0))) self.stopButton.SetPosition((marginHorizontal + 200, size[1] - marginBottom + max(100, size[0] / 15.0))) font = wx.Font(12, wx.FONTFAMILY_DEFAULT, wx.FONTSTYLE_NORMAL, wx.FONTWEIGHT_NORMAL) dc.SetTextForeground(colour) dc.SetFont(font) dc.DrawLabel('Input: %s, Output: %s' % (out.get_default_input_device_info()['name'], out.get_default_output_device_info()['name']), wx.Rect(marginHorizontal + 310, size[1] - marginBottom + 2 + max(100, size[0] / 15.0), 200, 100)) def OnClose(self, event): outputStream.stop_stream() outputStream.close() out.terminate() self.playing = False self.alive = False # self.recorder.join(1) self.Destroy() # exit() def OnDevice(self, event): # otherwise ask the user what new file to open with wx.FileDialog(self, "Open EQ .plist file", wildcard="plist files (*.plist)|*.plist", style=wx.FD_OPEN | wx.FD_FILE_MUST_EXIST) as fileDialog: if fileDialog.ShowModal() == wx.ID_CANCEL: return # the user changed their mind # Proceed loading the file chosen by the user pathname = fileDialog.GetPath() self.openDeviceFile(pathname) def openDeviceFile(self, pathname): global frequencies, interpolatedFrequencies, volumes, averageVolume, clipping try: frequencies = plistlib.readPlist(pathname) self.SetTitle(os.path.basename(os.path.splitext(pathname)[0])) interpolatedFrequencies = [] for i, f in enumerate(frequencies): if i > 0 and intermediateSteps > 0: for s in range(intermediateSteps): interpolatedFrequencies.append(Interpolate(frequencies[i-1], f, (s + 1)/float(intermediateSteps + 1))) interpolatedFrequencies.append(f) volumes = {} clipping = {} averageVolume = 0 for f in interpolatedFrequencies: volumes[f] = 0 self.Refresh() self.preferences.set('deviceFile', pathname) except IOError: wx.LogError("Cannot open file '%s'." % pathname) def OnPlay(self, event): if frequencies: self.playing = True def OnStop(self, event): if frequencies: self.currentFrequency = None self.Refresh() self.playing = False def DrawLine(self): dc = wx.ClientDC(self) dc.DrawLine(50, 60, 190, 60) if __name__ == '__main__': app = wx.App() e = Example(None, 'Frequency Response') e.DrawLine() e.Show() # e.play() e.DrawLine() app.MainLoop()
[ "post@yanone.de" ]
post@yanone.de
80d37d02b7af083195de59d9af45ee86907bdc6d
06651ac98fd2e89fa9a6b48fe536e5b93172909d
/object_decorator.py
e6dbf69910053c38780b152039108f16845707d1
[]
no_license
stephclleung/Worksheet-Generator
3c8e5d8d4e022c30073ea2d67074732074c3afc2
e3c5226db1ce949f1a2f22448b2bdadd493786f3
refs/heads/master
2020-04-12T10:14:38.948411
2019-02-25T07:34:43
2019-02-25T07:34:43
162,423,619
0
0
null
null
null
null
UTF-8
Python
false
false
2,559
py
# Objects - Decorator Pattern # # Stephanie Leung (2019) from random import randint from numpy.random import choice import nouns_gen as nouns_gen import adjective_gen as adj_gen DEBUG = False class AbstractObject: pass #Simplifying for now but may need to add in more later. class Noun(AbstractObject): def __init__(self, item_type): self.item_type = item_type self.noun = nouns_gen.noun_generate(self.item_type) def get_noun(self): return self.noun class AdjectiveDecorator(AbstractObject): def __init__(self, decorate_type): self.decorate_type = decorate_type def get_adj(self): if self.adj: return self.adj class State(AdjectiveDecorator): def __init__(self, decorate_type): AdjectiveDecorator.__init__(self, decorate_type) self.adj = adj_gen.state_adj_generator(decorate_type) class Quality(AdjectiveDecorator): def __init__(self, decorate_type): AdjectiveDecorator.__init__(self, decorate_type) self.adj = adj_gen.quality_adj_generator(decorate_type) class Size(AdjectiveDecorator): def __init__(self, decorate_type): AdjectiveDecorator.__init__(self, decorate_type) self.adj = adj_gen.size_adj_generator(decorate_type) class Color(AdjectiveDecorator): def __init__(self, decorate_type): AdjectiveDecorator.__init__(self, decorate_type) self.adj = adj_gen.color_adj_generator(decorate_type) def get_adjective(decorate_type): #adj_type = randint(1,3) # Color = 1 # Quality = 2 # Size = 3 # State = 4 if decorate_type in ["food", "drink", "appliance", "clothes"]: # all adjectives are fine adj_type = choice([1,2,3,4], p=[0.1, 0.2, 0.35, 0.35]) elif decorate_type == "activities": # quality and size only adj_type = choice([2,3], p=[0.65, 0.35]) elif decorate_type == "raw_mats": #color, quality only adj_type = choice([3,1], p=[0.35, 0.65]) if DEBUG: print(f"{decorate_type} {adj_type}") if adj_type == 1: return Color(decorate_type).get_adj() elif adj_type == 2: return Quality(decorate_type).get_adj() elif adj_type == 3: return Size(decorate_type).get_adj() elif adj_type == 4: return State(decorate_type).get_adj() def get_quantity_countable(decorate_type): return randint(1, 8) # 1/19 : Note on uncountable nouns, may need to come back and fix. # 1/19: Implementing temp fix on uncountable nouns (drinks, raw mats) def get_quantity_uncountable(decorate_type): return Quantity(decorate_type).get_adj() def get_noun(decorate_type): return Noun(decorate_type).get_noun() # test1 = Noun("food") # print(test1.noun) # print(set_adjective("food"))
[ "iLoveSalmon!1" ]
iLoveSalmon!1
5b17fbdae955c7e13c8c9c2642f0d53af96add6b
03f00d55672f0cd1edf2a5e5aacb4896aee624a9
/company/models.py
d28c471e7823cccb51433b71f60d03287a7ee2c9
[]
no_license
progiri/aiu_site
fe34cb884f01010e522855ade73c0a3be6476f02
17e81163c12d4861500d8d7d8e0207ded144f305
refs/heads/main
2023-06-11T15:04:50.658387
2021-06-29T04:17:17
2021-06-29T04:17:17
378,497,650
0
1
null
null
null
null
UTF-8
Python
false
false
1,524
py
from django.db import models class Location(models.Model): title = models.CharField( max_length=255, verbose_name='Название' ) class Meta: verbose_name = 'Местоположение (город)' verbose_name_plural = 'Местоположения (город)' def __str__(self) -> str: return self.title class Position(models.Model): title = models.CharField( max_length=255, verbose_name='Название' ) class Meta: verbose_name = 'Должность' verbose_name_plural = 'Должности' def __str__(self) -> str: return self.title class Company(models.Model): title = models.CharField( max_length=255, verbose_name='Название' ) description = models.CharField( max_length=2555, verbose_name='Описание' ) location = models.ForeignKey( to=Location, on_delete=models.SET_NULL, related_name='companies', null=True, blank=True, verbose_name='Местоположение (город)' ) position = models.ForeignKey( to=Position, on_delete=models.SET_NULL, related_name='companies', null=True, blank=True, verbose_name='Должность' ) class Meta: verbose_name = 'Компания' verbose_name_plural = 'Компании' def __str__(self) -> str: return self.title
[ "yerassyl.ak@mail.ru" ]
yerassyl.ak@mail.ru
cdede1a5eecbcfedafb5177344e00ab2dc3e9821
a65e5dc54092a318fc469543c3b96f6699d0c60b
/Personel/Yash/Python/april5/prog5.py
a9186b792a3f46d892bdf941b3d00acb7fecf2ff
[]
no_license
shankar7791/MI-10-DevOps
e15bfda460ffd0afce63274f2f430445d04261fe
f0b9e8c5be7b28298eb6d3fb6badf11cd033881d
refs/heads/main
2023-07-04T15:25:08.673757
2021-08-12T09:12:37
2021-08-12T09:12:37
339,016,230
1
0
null
2021-08-12T09:12:37
2021-02-15T08:50:08
JavaScript
UTF-8
Python
false
false
281
py
# To delete a file, you must import the OS module, and run its os.remove() function: import os if os.path.exists("demofile.txt"): os.remove("demofile.txt") else: print("The file does not exist") # To delete an entire folder, use the os.rmdir() method: # os.rmdir("myfolder")
[ "malavade47@gmail.com" ]
malavade47@gmail.com
adfd6bfcac5e5eed8bffd7e1f79d115c31bb2607
da04aa71c2802fa6f9c1fd463f31ad4e4e514df9
/heartbeat plot.py
dba47c3bd4c508f9c1ad21e9e5c136d7508d4c77
[]
no_license
JayDosunmu/shield
e698dde7de57277d361e212c41ea986c0c7f21ad
89f9444f6f98f06db0c04858f1b2bee1ce01ca43
refs/heads/master
2021-01-17T09:46:29.085320
2017-03-05T18:52:47
2017-03-05T18:52:47
83,992,746
1
0
null
null
null
null
UTF-8
Python
false
false
509
py
import plotly.plotly as py import plotly.graph_objs as go import pyrebase import time config = { "apiKey":"AlzaSyBcPkjp50hVFQj3jL7NdCMul0Cw9jP5gkc", "authDomain":"hacktech-12dad.firebaseapp.com", "databaseURL":"https://hacktech-12dad.firebaseio.com", "storageBucket":"", } fb = pyrebase.initialize_app(config) db = fb.database() data = db.child('snapshots').get() data = data.each() print(data) y = [change.get('heartbeat').get('change') for change in data] #print(y)
[ "tjdosunmu@gmail.com" ]
tjdosunmu@gmail.com
56255c3a178b2823db3fdb697e972ac01d1e2f03
3a5dab20e575c9d6d1f671faba7da8f6085d9ed8
/RepositorioInfo/chacoferia/apps/inicio/views.py
eab1b18c67ddcafd6cb4c9da16ca566ff43dd7c6
[]
no_license
Diego-Caza/ProyectoFinal
77cb45310375e31c2924b10ecd2dfad9533f2982
47ce682027d4230a494e4b875367a65e386a0746
refs/heads/master
2022-12-19T12:35:49.020185
2020-10-01T16:37:31
2020-10-01T16:37:31
300,335,385
1
1
null
null
null
null
UTF-8
Python
false
false
455
py
from django.shortcuts import render from apps.publicaciones.models import Producto from django.views.generic import ListView, DetailView, CreateView # Create your views here. class Populares(ListView): model = Producto template_name = 'inicio/inicio.html' queryset = Producto.objects.order_by('-id')[:4] class PopularesUser(ListView): model = Producto template_name = 'inicio/usuario/inicio.html' queryset = Producto.objects.order_by('-id')[:4]
[ "cazadiegoy@gmail.com" ]
cazadiegoy@gmail.com
e9a4c0ee8774a16092863b3972e7e903593cac32
492cb86b533bc74962a0e25ad190dab131f7cb09
/humanScape/urls.py
d66fdef307d8290976f7ee67668986092280f3c9
[]
no_license
acdacd66/humanscape
75f27815f6c1ac5975b3822e5abc5738aa9b3118
6fbeeca3346569c7f861bbffcbec731a6a9d6e51
refs/heads/main
2023-09-02T01:55:49.806746
2021-11-16T17:29:36
2021-11-16T17:29:36
428,570,173
0
1
null
2021-11-16T11:22:32
2021-11-16T08:10:30
Python
UTF-8
Python
false
false
820
py
"""humanScape URL Configuration The `urlpatterns` list routes URLs to views. For more information please see: https://docs.djangoproject.com/en/3.2/topics/http/urls/ Examples: Function views 1. Add an import: from my_app import views 2. Add a URL to urlpatterns: path('', views.home, name='home') Class-based views 1. Add an import: from other_app.views import Home 2. Add a URL to urlpatterns: path('', Home.as_view(), name='home') Including another URLconf 1. Import the include() function: from django.urls import include, path 2. Add a URL to urlpatterns: path('blog/', include('blog.urls')) """ from django.contrib import admin from django.urls import path,include urlpatterns = [ path('admin/', admin.site.urls), path("clinical/", include("clinicalInformation.urls")), ]
[ "you@example.com" ]
you@example.com
f73086d0105b3dbd3762dcdd037934533567ee3a
3f8ef6ce1c878e6aa358714ea3c85a988de8948f
/Evolution Software Engineer/tenth_year.py
c97bbc78b102ba361ad8c494668ff7c884f5cee6
[]
no_license
gchacaltana/PythonToolkit
1ae320bb091996a5214fc446345f115cdcb28202
05d683fb5170294e95f4d42ed5bbe96135e9538e
refs/heads/master
2022-03-08T02:47:04.823608
2022-02-11T23:11:22
2022-02-11T23:11:22
24,420,758
0
0
null
null
null
null
UTF-8
Python
false
false
271
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """Programa que imprime en pantalla el mensaje Hola Mundo""" __author__ = 'Gonzalo Chacaltana' __email__ = "gchacaltanab@outlook.com" class HelloWorld(object): def __init__(self): print ("Hola Mundo") obj = HelloWorld()
[ "gchacaltanab@gmail.com" ]
gchacaltanab@gmail.com
1aaceb54e5b237cd1f0061a9184ee10f4afd7a63
cba5017525d30f84f4555bc0e10f1f83126f1d4a
/PowerClient/bin/main.py
2b7af5a8bdb8d84c2687ddb6d2eed2ca8d1375e7
[ "Apache-2.0" ]
permissive
cycmay/SolarS
66e97a0de6b459f8bb05b03c2690d9852d92209a
284bcafa5da210e5c4200d19e46b3fa6bb5acb20
refs/heads/master
2020-05-23T18:09:09.760666
2019-05-24T15:42:25
2019-05-24T15:42:25
186,882,616
0
0
null
null
null
null
UTF-8
Python
false
false
392
py
#!/usr/bin/env python # -*- coding:utf-8 -*- """ 完全可以把信息收集客户端做成Windows和Linux两个不同版本。 """ import os import sys BASE_DIR = os.path.dirname(os.getcwd()) # 设置工作目录,使得包和模块能够正常导入 sys.path.append(BASE_DIR) from core import handler if __name__ == '__main__': handler.ArgvHandler(sys.argv)
[ "1769614470@qq.com" ]
1769614470@qq.com
3daab6c956e8d126316ecdb6ef6e71d8af6a258d
1c8a1b7cfb5c78fe94c4cc62a78dbfff96161924
/day05/test04.py
7715b05a49b005d9cad71dc19124fa6797945c72
[]
no_license
WHUTyuen/PIL_opencv
d264858f0eaa4ecc555747efd5f277f48a432b91
3ae6e7d878215866c304e64eac05bf1011ecb428
refs/heads/main
2023-01-01T14:00:33.331676
2020-11-01T11:35:18
2020-11-01T11:35:18
309,072,309
0
0
null
null
null
null
UTF-8
Python
false
false
1,122
py
import cv2 import numpy as np A = cv2.imread('3.jpg') B = cv2.imread('4.jpg') G = A.copy() gpA = [G] for i in range(6): G = cv2.pyrDown(G) gpA.append(G) G = B.copy() gpB = [G] for i in range(6): G = cv2.pyrDown(G) gpB.append(G) # generate Laplacian Pyramid for A lpA = [gpA[5]] for i in range(5, 0, -1): GE = cv2.pyrUp(gpA[i]) L = cv2.subtract(gpA[i - 1], GE) lpA.append(L) # generate Laplacian Pyramid for B lpB = [gpB[5]] for i in range(5, 0, -1): GE = cv2.pyrUp(gpB[i]) L = cv2.subtract(gpB[i - 1], GE) lpB.append(L) # Now add left and right halves of images in each level LS = [] for la, lb in zip(lpA, lpB): rows, cols, dpt = la.shape ls = np.hstack((la[:, 0:cols // 2], lb[:, cols // 2:])) LS.append(ls) # now reconstruct ls_ = LS[0] for i in range(1, 6): ls_ = cv2.pyrUp(ls_) ls_ = cv2.add(ls_, LS[i]) # image with direct connecting each half real = np.hstack((A[:, :cols // 2], B[:, cols // 2:])) cv2.imshow('Pyramid_blending.jpg', ls_) cv2.imshow('Direct_blending.jpg', real) cv2.waitKey(0)
[ "noreply@github.com" ]
noreply@github.com
79feeaca9b02f648fa48495c3481b1b1940c3057
406abf223c6eaca3ec4741e7136cfc2fe819022b
/api/utils.py
d778f462a585c6d1f73e13507c6ed37c3e75ed4a
[]
no_license
WarrentLee/ImgChangeServer
e068132e2430863698870038d2627a79e2bb51b3
14e9658e75ab2eebe0cbafa119410ec2c6200321
refs/heads/master
2022-10-07T15:36:03.136973
2020-06-10T01:41:50
2020-06-10T01:41:50
269,850,341
0
0
null
null
null
null
UTF-8
Python
false
false
636
py
from flask import Flask from database.model import db import os, json from config import Config def create_app(name): app = Flask(name, template_folder='../templates') # app.config['SQLALCHEMY_DATABASE_URI'] = os.getenv('DATABASE_URL', 'sqlite:///' + os.path.join(app.root_path, 'data.db')) print("Database Directory:", os.getenv('DATABASE_URL')) app.config['SQLALCHEMY_DATABASE_URI'] = os.getenv('DATABASE_URL', 'sqlite:///' + Config.DATABASE_DIRECTORY) app.config['SQLALCHEMY_TRACK_MODIFICATIONS'] = False app.secret_key = "asfnsdjnsdlflasdkc553d3s1" db.app = app db.init_app(app) return app
[ "1065791740@qq.com" ]
1065791740@qq.com
868e5aecbeb632ef54b45e30b589b06ef555afab
8a8dc992231e30f70e2eb645a2d869a730d3c052
/backendvfinal4.py
c2f42ef346ee5a8a29839e1e19c4f4175b16f97a
[]
no_license
severlove/backend_project
111c3f34fb9891f6a1e0f02cab6530e099215dca
48c3f2c0d2c7ff6bd5d520303eda4deff229611e
refs/heads/master
2020-04-27T07:53:17.313968
2019-03-07T16:37:09
2019-03-07T16:37:09
174,151,363
0
0
null
null
null
null
UTF-8
Python
false
false
2,209
py
#!/usr/bin/env python3 import psycopg2 def top_3_articles(): '''Print top 3 articles of all time''' db = psycopg2.connect("dbname=news") c = db.cursor() query = """SELECT articles.title, count(articles.title) as num FROM articles,log WHERE articles.slug = SUBSTRING(log.path, 10) GROUP BY title ORDER BY num DESC LIMIT 3;""" c.execute(query) results = c.fetchall() db.close() print("\nTop 3 Articles:\n") for i in range(len(results)): print "\"" + results[i][0] + "\" - " + str(results[i][1]) + " views" def popular_authors(): '''Print the top authors''' db = psycopg2.connect("dbname=news") c = db.cursor() query = """SELECT authors.name, count(articles.author) as num FROM articles, log, authors WHERE articles.slug = SUBSTRING(log.path, 10) and articles.author = authors.id GROUP BY authors.name ORDER BY num DESC""" c.execute(query) results = c.fetchall() db.close() print("\nPopular Authors:\n") for i in range(len(results)): print results[i][0] + " - " + str(results[i][1]) + " views" def errors(): '''Print the date where more than 1% of requests cause errors''' db = psycopg2.connect("dbname=news") c = db.cursor() query = """SELECT Date, Total, Error, (Error::float*100)/Total::float as Percent FROM ( SELECT to_char(time::timestamp::date, 'Month DD, YYYY') as Date, count(status) as Total, sum(case when not status = '200 OK' then 1 else 0 end) as Error FROM log GROUP BY time::timestamp::date) as result WHERE(Error::float*100)/Total::float > 1.0;""" c.execute(query) results = c.fetchall() db.close() print("\nDays where more than 1% of requests lead to errors:\n") for i in range(len(results)): print(str(results[i][0]) + " - " + str(round(results[i][3], 1)) + "% errors") print("\n") if __name__ == '__main__': top_3_articles() popular_authors() errors()
[ "noreply@github.com" ]
noreply@github.com
3f17973dce3bd98e0003d899e9f727b81dc0574e
27767ea4ec9cb4b3d8a3ba31c9e668957d3f900c
/tools/statistic.py
4606f091b46a67087b07c966d225389bbf7ff6f9
[]
no_license
01ghost13/smartyPantsBot
8609dd71bfaf0ace1d137cc7bbe4cf87e6b9f1b6
bc8943b416e42be263cc617c17e702b978aeab8b
refs/heads/master
2020-04-10T07:27:03.989421
2019-05-12T12:55:38
2019-05-12T13:04:55
160,880,698
0
0
null
2018-12-16T14:52:15
2018-12-07T22:12:11
Python
UTF-8
Python
false
false
666
py
from tools import Database class Statistic: smile_count_sql = """ select e.name, e.discord_id, count(e_m.id) as smile_count from emojies as e join emoji_messages as e_m on e_m.emoji_id = e.id join messages as m on m.id = e_m.message_id where m.channel_id = '%s' group by e.name, e.discord_id order by smile_count desc """ @staticmethod def smile_count(channel_id): with Database() as db: conn = db.connect() cursor = conn.cursor() cursor.execute(Statistic.smile_count_sql % channel_id) return cursor.fetchall()
[ "orkich@gmail.com" ]
orkich@gmail.com
38840ebcb8cb679446cceb6b8beace276bf9b1a7
415e9349304a1fc2555a55ff0e3caff3fce26e68
/hydrogen_atom.py
214b1a0a8884d76d855e9fc4a7fc2faa8d7bf7eb
[]
no_license
einathauser/tau-dmc-2017
9a3ba554e5565628c227c8ba04167ef8790f6571
154a953b5905545739908f4216e33b2204c151ff
refs/heads/master
2021-01-23T09:39:49.761259
2017-09-13T13:12:54
2017-09-13T13:12:54
102,589,773
0
0
null
null
null
null
UTF-8
Python
false
false
3,560
py
import numpy as np import matplotlib.pyplot as plt from matplotlib.animation import FuncAnimation import matplotlib.mlab as mlab MAX_X_SIZE = 1e6 def sample_bins(x_max, x, n_boxes): r_max = np.sqrt(float(3*x_max**2)) radius = np.array(np.sqrt(x[:, 0]**2 + x[:, 1]**2 + x[:, 2]**2)).astype(float) bins, _ = np.histogram(radius, bins=np.linspace(0.0, r_max, n_boxes + 1)) return bins def wave_function(density, x_max, n_boxes, bin_size): sum_of_square = 0 delta_r = np.sqrt(3*x_max**2).astype(float)/n_boxes new_matrix = (density**2) for item in new_matrix: sum_of_square += item psi_x = density.astype(float) / ((bin_size)*float(np.sqrt(delta_r*(sum_of_square)))) return psi_x def m(W_x): return np.minimum(np.trunc(W_x + np.random.uniform(0.0, 1.0, W_x.size)), 3).astype(int) def V(e, x): distance = np.array(np.sqrt(x[:, 0]**2 + x[:, 1]**2 + x[:, 2]**2)).astype(float) return ((-e**2) / distance) def W(V_x, e_r, dt): return np.exp(-(V_x - e_r) * dt) def particle_locations(x, dt): sigma = np.sqrt(dt) row = np.random.randn(len(x), 3) location = x + sigma*row return location def energy(v_x, n_pre, n_0, dt): return np.average(v_x) + (1.0 - float(n_pre) / n_0) / dt # if (previous * 100 / avg_e_r) <= 5: # time_interval.append(time) def run_dmc(dt, n_times, n_0, x_min, x_max, n_boxes, sample_from_iteration, e): x = np.array([[0, 0, 1]]*n_0).astype(float) V_x = V(e, x) e_r = np.average(V_x) e_rs = [e_r] bins = np.zeros(n_boxes) psi = 0 a = np.arange(n_boxes) bin_size = (np.sqrt(3*x_max**2).astype(float)/n_boxes)*(a + 0.5) matrix_of_volumes = (0+(4*np.pi*(float(np.sqrt(3*(x_max**2)))**3) / (3*n_boxes**3)))*(a + 0.5) for i in range(n_times): # creates a vector of number with step dt. i gives the items in the list x = particle_locations(x, dt) V_x = V(e,x) W_x = W(V_x, e_r, dt) m_x = m(W_x) x = np.repeat(x, m_x, axis=0) n_previous = len(x) # print('Round %d m_x: %s' % (i, np.mean(m_x))) e_r = energy(V_x, n_previous, n_0, dt) print e_r # previous_avg = np.average(e_rs) e_rs.append(e_r) if len(x) > MAX_X_SIZE: raise Exception('x is too big, aborting!') if i > sample_from_iteration: bins += sample_bins(x_max, x, n_boxes) density = bins/matrix_of_volumes psi = wave_function(density, x_max, n_boxes, bin_size) avg_e_r = np.average(e_rs[sample_from_iteration:]) standard_dev = np.std(e_rs[sample_from_iteration:]) r_psi = psi*bin_size plt.xlim(0.0, 7.5) plt.ylim(0.0, 2.2) plt.title("DMC Hydrogen atom") psi_analytic = 2*np.exp(-bin_size) r_psi_analytic = bin_size*psi_analytic plt.plot(bin_size, r_psi, color = 'green') # plt.plot(e_rs) plt.plot(bin_size, psi_analytic, color = 'blue') plt.plot(bin_size, psi, color = 'red') plt.plot(bin_size, r_psi_analytic, color='yellow') plt.show() return standard_dev, avg_e_r if __name__ == "__main__": # execute only if run as a script n_0 = 5000 x_min = 0.0 x_max = 10.0 # x_max=y_max=z_max. these are the values which indicate location that you can enter the matrix # the values can be different but their range can't because it's a sphere. n_boxes = 2000 dt = 0.1 n_times = 2000 sample_from_iteration = 100 e = 1 print run_dmc(dt, n_times, n_0, x_min, x_max, n_boxes, sample_from_iteration, e)
[ "ehauzer11@gmail.com" ]
ehauzer11@gmail.com
072124f7bf3fc900a33e0ea67e58898b947b9d5c
87998e751e753dd4d75848ee292631d96b71d608
/aoc/year2020/day2.py
c37501a97f7ff8492b062238bfec7e3aaeb5d164
[ "Unlicense" ]
permissive
ongyx/aoc.py
7da09592985c6688fb766f6c4b80526f51d2c290
48d23b2611384cf87b2722d6fd65728bb950c3f9
refs/heads/main
2023-01-31T07:50:44.349391
2020-12-16T03:43:06
2020-12-16T03:43:06
null
0
0
null
null
null
null
UTF-8
Python
false
false
934
py
# coding: utf8 import re from aoc import aoc # We use a regex to parse each line, in the format: # (min)-(max) (letter): (password) # For part 2, we need to fidde with the Boolean logic abit. RE_POLICY = re.compile(r"(\d+)-(\d+) (\w): (\w+)") class Problem(aoc.Problem): def solve(self, part): vaild_counter = 0 for line in self.dataset_lines: lower, upper, letter, password = RE_POLICY.findall(line)[0] lower = int(lower) upper = int(upper) if part == 1: if lower <= password.count(letter) <= upper: vaild_counter += 1 elif part == 2: # only one of the positions can have the letter # so we have to use a XOR (exclusive OR). if (password[lower - 1] == letter) ^ (password[upper - 1] == letter): vaild_counter += 1 return vaild_counter
[ "ongyongxin2020@gmail.com" ]
ongyongxin2020@gmail.com
98872d68f0a8214b87de5d704fb379cb99c72671
6f76052e2a1161b8a728dd6fff6dfafd107c6c57
/Kuznechik.py
58c46f3b1ea4e3f98d1464f74da175e13cfeeead
[]
no_license
mratnikov/Rain
e33ed9bf51a0eebd8a4cd1e8c6f039fc511658bf
a11c662a87647078141fb213379c544b258881fc
refs/heads/master
2020-03-24T11:59:53.650407
2019-03-14T17:38:23
2019-03-14T17:38:23
142,700,601
1
0
null
null
null
null
UTF-8
Python
false
false
100
py
F = [0] * (n + 1) F[0] = 1 F[1] = 1 for i in range(2, n + 1): F[i] = F[i - 2] + F[i — 1]
[ "mratnikov@gmail.com" ]
mratnikov@gmail.com
f25d084d56d88bcd33d43a1a0ea324541f6d799f
ae9ff58a96682e57da9b6a1376db7ece5d84baa2
/contrib/bulk_operations/bulk_operations.py
005659e1475c3241682c87ea583fa35188da1cdc
[ "MIT" ]
permissive
simozhan/product-definition-center
0ca1fde76c9c80368e3116dd68465fe91a290c7f
55c2de179cf4169213c8a8bd81977e0cdc8ee350
refs/heads/master
2021-01-22T16:05:30.731848
2015-07-24T04:46:16
2015-07-24T04:46:16
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,001
py
# # Copyright (c) 2015 Red Hat # Licensed under The MIT License (MIT) # http://opensource.org/licenses/MIT # """ This module provides BulkRouter that extends the registered ViewSets with bulk operations if they are not provided yet. To display documentation in the browsable API, it is necessary to provide a method `bulk_op` (where `op` is any of `update`, `partial_update`, `destroy`) on the viewset which calls `pdc.apps.common.bulk.bulk_op_impl` instead of a parent class. Bulk create does not get its own tab in browsable API. If the docstrings for these methods are not provided, they come with some generic ones. """ from functools import wraps from collections import OrderedDict from rest_framework.settings import api_settings from rest_framework import routers, status from rest_framework.response import Response from django.conf import settings def _failure_response(ident, response, data=None): """ Given an identifier, a response from a view and optional data, return a Response object that describes the error. """ result = { 'invalid_data_id': ident, 'detail': response.data.get('detail', response.data), } if data: result['invalid_data'] = data response = Response(result, status=response.status_code) # This tells ChangesetMiddleware to abort the transaction. response.exception = True return response def _safe_run(func, *args, **kwargs): """ Try to run a function with given arguments. If it raises an exception, try to convert it to response with the exception handler. If that fails, the exception is re-raised. """ try: return func(*args, **kwargs) except Exception, exc: response = api_settings.EXCEPTION_HANDLER(exc, context=kwargs) if response is not None: return response raise def bulk_create_wrapper(func): @wraps(func) def wrapper(self, request, *args, **kwargs): data = request.data if not isinstance(data, list): return func(self, request, *args, **kwargs) result = [] for idx, obj in enumerate(data): request._full_data = obj response = _safe_run(func, self, request, *args, **kwargs) if not status.is_success(response.status_code): return _failure_response(idx, response, data=obj) # Reset object in view set. setattr(self, 'object', None) result.append(response.data) return Response(result, status=status.HTTP_201_CREATED) return wrapper def bulk_destroy_impl(self, request, **kwargs): """ It is possible to delete multiple items in one request. Use the `DELETE` method with the same url as for listing/creating objects. The request body should contain a list with identifiers for objects to be deleted. The identifier is usually the last part of the URL for deleting a single object. """ if not isinstance(request.data, list): return Response(status=status.HTTP_400_BAD_REQUEST, data={'detail': 'Bulk delete needs a list of identifiers.'}) self.kwargs.update(kwargs) for ident in OrderedDict.fromkeys(request.data): self.kwargs[self.lookup_field] = unicode(ident) response = _safe_run(self.destroy, request, **self.kwargs) if not status.is_success(response.status_code): return _failure_response(ident, response) return Response(status=status.HTTP_204_NO_CONTENT) def bulk_update_impl(self, request, **kwargs): """ It is possible to update multiple objects in one request. Use the `PUT` or `PATCH` method with the same url as for listing/creating objects. The request body should contain an object, where keys are identifiers of objects to be modified and their values use the same format as normal *update*. """ if not isinstance(request.data, dict): return Response(status=status.HTTP_400_BAD_REQUEST, data={'detail': 'Bulk update needs a mapping.'}) result = {} self.kwargs.update(kwargs) orig_data = request.data for ident, data in orig_data.iteritems(): self.kwargs[self.lookup_field] = unicode(ident) request._full_data = data response = _safe_run(self.update, request, **self.kwargs) if not status.is_success(response.status_code): return _failure_response(ident, response, data=data) result[ident] = response.data return Response(status=status.HTTP_200_OK, data=result) def bulk_partial_update_impl(self, request, **kwargs): if not request.data: return Response( status=status.HTTP_400_BAD_REQUEST, data=settings.EMPTY_PATCH_ERROR_RESPONSE ) self.kwargs['partial'] = True return self.bulk_update(request, **kwargs) def bulk_create_dummy_impl(): """ It is possible to create this resource in bulk. To do so, use the same procedure as when creating a single instance, only the request body should contain a list of JSON objects. The response you get back will also contain a list of values which you would obtain by submitting the request data separately. """ assert False, ('This method should never be called, it is here just so ' 'that there is a method to attach a docstring to.') class BulkRouter(routers.DefaultRouter): """ This router provides the standard set of resources (the same as `DefaultRouter`). In addition to that, it allows for bulk operations on the collection as a whole. These are performed as a PUT/PATCH/DELETE request on the `{basename}-list` url. These requests are dispatched to the `bulk_update`, `bulk_partial_update` and `bulk_destroy` methods respectively. The bulk create does not have a dedicated method (because the URL and method are the same as for regular create). Currently, there is no way to opt-out from having bulk create added. It is however possible to define a method named `bulk_create` which will provide docstring to be rendered in browsable API. This method will never be called. If the method is missing, a generic documentation will be added. """ def get_routes(self, viewset): for route in self.routes: if isinstance(route, routers.Route) and route.name.endswith('-list'): route.mapping.update({'delete': 'bulk_destroy', 'put': 'bulk_update', 'patch': 'bulk_partial_update'}) return super(BulkRouter, self).get_routes(viewset) def register(self, prefix, viewset, base_name=None): if hasattr(viewset, 'create'): viewset.create = bulk_create_wrapper(viewset.create) if not hasattr(viewset, 'bulk_create'): viewset.bulk_create = bulk_create_dummy_impl if hasattr(viewset, 'destroy') and not hasattr(viewset, 'bulk_destroy'): viewset.bulk_destroy = bulk_destroy_impl if hasattr(viewset, 'update') and not hasattr(viewset, 'bulk_update'): viewset.bulk_update = bulk_update_impl if hasattr(viewset, 'partial_update') and not hasattr(viewset, 'bulk_partial_update'): viewset.bulk_partial_update = bulk_partial_update_impl super(BulkRouter, self).register(prefix, viewset, base_name) def get_lookup_regex(self, viewset, lookup_prefix=''): """ For viewsets using the MultiLookupFieldMixin, it is necessary to construct the lookup_value_regex attribute here. """ if hasattr(viewset, 'lookup_fields'): regexes = [] for field_name, field_regex in viewset.lookup_fields: regexes.append('(?P<%s>%s)' % (field_name, field_regex)) viewset.lookup_value_regex = '/'.join(regexes) return super(BulkRouter, self).get_lookup_regex(viewset, lookup_prefix)
[ "lsedlar@redhat.com" ]
lsedlar@redhat.com
4cf93fc1cb9c2a0a2877f9616693f8488607baa0
10e8c5f2f917fd091b0bab7aed18f4130039fa86
/day_7_2.py
71299c297d0081f51e139cc688aec2245d0f6f76
[]
no_license
DannyBarbaro/AOC_2019
9383a2f14c50cbed9c359fff5f62edd291e1da77
e1b7948ebc76af9e5bab076c18a6140e54e54f64
refs/heads/master
2020-12-03T17:56:23.228622
2020-01-02T16:30:14
2020-01-02T16:30:14
231,418,989
0
0
null
null
null
null
UTF-8
Python
false
false
3,312
py
datastream = [] def get_value(val, mode, data): return data[val] if mode == 0 else val def oppcode_1(val_1, mode_1, val_2, mode_2, result, data): data[result] = get_value(val_1, mode_1, data) + get_value(val_2, mode_2, data) return data def oppcode_2(val_1, mode_1, val_2, mode_2, result, data): data[result] = get_value(val_1, mode_1, data) * get_value(val_2, mode_2, data) return data def oppcode_3(val, data): data[val] = datastream.pop(0) return data def oppcode_4(val, mode, data): return get_value(val, mode, data) def oppcode_5(val_1, mode_1, data): return True if get_value(val_1, mode_1, data) != 0 else False def oppcode_6(val_1, mode_1, data): return True if get_value(val_1, mode_1, data) == 0 else False def oppcode_7(val_1, mode_1, val_2, mode_2, val_3, data): num = 1 if get_value(val_1, mode_1, data) < get_value(val_2, mode_2, data) else 0 data[val_3] = num return data def oppcode_8(val_1, mode_1, val_2, mode_2, val_3, data): num = 1 if get_value(val_1, mode_1, data) == get_value(val_2, mode_2, data) else 0 data[val_3] = num return data def run_intcode(ints, index): while True: code = ints[index] op = code % 100 mode_1 = code % 1000 - code % 100 mode_2 = code % 10000 - code % 1000 mode_3 = code % 100000 - code % 10000 if op == 1: ints = oppcode_1(ints[index+1], mode_1, ints[index+2], mode_2, ints[index+3], ints) index += 4 elif op == 2: ints = oppcode_2(ints[index+1], mode_1, ints[index+2], mode_2, ints[index+3], ints) index += 4 elif op == 3: ints = oppcode_3(ints[index+1], ints) index += 2 elif op == 4: return [oppcode_4(ints[index+1], mode_1, ints), ints, index+2] elif op == 5: if oppcode_5(ints[index+1], mode_1, ints): index = get_value(ints[index+2], mode_2, ints) else: index += 3 elif op == 6: if oppcode_6(ints[index+1], mode_1, ints): index = get_value(ints[index+2], mode_2, ints) else: index += 3 elif op == 7: ints = oppcode_7(ints[index+1], mode_1, ints[index+2], mode_2, ints[index+3], ints) index += 4 elif op == 8: ints = oppcode_8(ints[index+1], mode_1, ints[index+2], mode_2, ints[index+3], ints) index += 4 elif op == 99: return def new_state(): ints = [] with open('inputs/input7.txt') as f: return [int(numeric_string) for numeric_string in f.readline().split(',')] def calc_for_perm(combo): datastream.clear() states = [[],[],[],[],[]] for i in range(len(states)): states[i] = new_state() pointers = [0,0,0,0,0] val = 0 isFirst = True running = True while running: for i in range(len(combo)): if isFirst: datastream.append(combo[i]) datastream.append(val) output = run_intcode(states[i], pointers[i]) if output == None: running = False else: val = output[0] states[i] = output[1] pointers[i] = output[2] isFirst = False return val # Part 2 import itertools import time perms = list(itertools.permutations([5,6,7,8,9])) maximum = 0 best = [] for combo in perms: combo = list(combo) val = calc_for_perm(combo) if val > maximum: maximum = val best = combo print(maximum) print(combo) # expected = 139629729
[ "dxb486@case.edu" ]
dxb486@case.edu
27c38c01ec059532373e8cd03289ccde4ded2e1d
f0f3f8731145e236e8e08dafb4201108d35af488
/wish_list_items/migrations/0007_auto_20160414_1317.py
8478f0d9cfbb5235617279dac1587637337832db
[]
no_license
AaronScruggs/wish_list_project
49fdfc9c3a9e72470084bbf283085c15aa659a3e
a2a741823e0a570390ce344f3407f6f3b57f2590
refs/heads/master
2021-01-01T05:18:10.817456
2016-04-19T00:36:24
2016-04-19T00:36:24
56,259,190
0
0
null
null
null
null
UTF-8
Python
false
false
471
py
# -*- coding: utf-8 -*- # Generated by Django 1.9.5 on 2016-04-14 20:17 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('wish_list_items', '0006_auto_20160414_1312'), ] operations = [ migrations.AlterField( model_name='wishitem', name='item_url', field=models.URLField(default=True, null=True), ), ]
[ "aarondscruggs@gmail.com" ]
aarondscruggs@gmail.com
276ee2323e5bd2b36459c096005c556b3732ee1f
6832a7fa70fc98f62b777421fca13989b8e2e1e9
/greeting_decorator.py
6254f95fa5c91f95649dcbfae0570d84bbd5cc75
[]
no_license
serdarayalp/my-python
f553df88b284528dddd245557935f349b34df963
1e23e6b9cccdeb96f3bfd9a0812638dd8c02c189
refs/heads/main
2023-02-28T00:39:24.691917
2021-02-06T08:53:30
2021-02-06T08:53:30
330,440,789
0
0
null
null
null
null
UTF-8
Python
false
false
244
py
from functools import wraps def greeting(func): @wraps(func) def function_wrapper(x): """ function_wrapper of greeting """ print("Hi, " + func.__name__ + " returns:") return func(x) return function_wrapper
[ "serdarayalp@googlemail.com" ]
serdarayalp@googlemail.com
8104871522b317d80e9b8bcd58d1d04195096050
57385c8ae2b10d01c2dede1d69f1887d8a8a7ac0
/manage.py
001c29815eafae252644ec0a25538483b622b402
[ "MIT" ]
permissive
bruno-dash/teste-5
151efb6558168c20d3320c510893a922e2d68ee1
484cc065673fe859cda571b208be79d82368eeae
refs/heads/main
2023-08-19T01:37:10.334089
2021-10-16T23:27:45
2021-10-16T23:27:45
417,916,703
0
0
null
null
null
null
UTF-8
Python
false
false
662
py
#!/usr/bin/env python """Django's command-line utility for administrative tasks.""" import os import sys def main(): """Run administrative tasks.""" os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'teste5.settings') try: from django.core.management import execute_from_command_line except ImportError as exc: raise ImportError( "Couldn't import Django. Are you sure it's installed and " "available on your PYTHONPATH environment variable? Did you " "forget to activate a virtual environment?" ) from exc execute_from_command_line(sys.argv) if __name__ == '__main__': main()
[ "bonofre36@gmail.com" ]
bonofre36@gmail.com
8424d48df45764507a31d8a1ab65bd44b3544d9b
aa2210cbc53e2892f2d7192085433950879b3e93
/file_summary_1.py
a2aa31b77215d790056587eed7df6145cb8b9d87
[]
no_license
6188506/LearnPythonHardWay
dddc9f076902e94ba1cbff24142237ea7cdeeb28
bd071a6602a1cde297209fa2339ebe3d96da6441
refs/heads/master
2021-01-20T04:11:36.536522
2018-05-03T09:01:38
2018-05-03T09:01:38
89,654,909
1
0
null
null
null
null
UTF-8
Python
false
false
202
py
#from sys import argv #script, destination_file = argv #print script #print "read file from ex1 to ex41,and write to a argv file" x = 6 print open('e:/learnpythonhardway/ex'+str(x)+'.py','r').read()
[ "6188506@qq.com" ]
6188506@qq.com
3d4eeac4dc2bacb190a2ec396de0cfce8de00b15
902215df8f073b83b5815657d7b69fd6d67c8b5f
/kkd/views.py
00c6807b8480442a788fd0a2ca628a9bf47f4ca8
[]
no_license
k821209/KKD_GO_DB
e7643209da7d33cc1eee1395430046fd41199af7
9745dcb2c2a62db08b239716dfc4f3581200bb8c
refs/heads/master
2020-04-07T05:48:01.054782
2018-05-02T05:48:47
2018-05-02T06:33:00
124,186,255
0
0
null
null
null
null
UTF-8
Python
false
false
143
py
from django.shortcuts import render def show_go_list(request): return render(request, 'kkd/go_index.html',{}) # Create your views here.
[ "kangyangjae@gmail.com" ]
kangyangjae@gmail.com
d3d12734a381e7185f0368cdd0b926088c0b7116
648f5819d70686852720d6b598b66ea37cc49b71
/setup.py
1a9e1c6cd8a870c6d9ed3796176d7514507caf98
[ "MIT" ]
permissive
itoldman/chinese_province_city_area_mapper
88981f1e74507ac3cdd8ff0b60a7b5716bb701f4
7d7537a412dc1c543f3481f5520b3c75cf49ebb0
refs/heads/master
2020-03-30T12:46:11.678976
2018-09-26T17:59:50
2018-09-26T17:59:50
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,033
py
# -*- coding: utf-8 -*- from setuptools import setup LONGDOC = """ chinese_province_city_area_mapper ================================== chinese_province_city_area_mapper:一个用于识别简体中文字符串中省,市和区并能够进行映射,检验和简单绘图的python模块 举个例子:: ["徐汇区虹漕路461号58号楼5楼", "泉州市洛江区万安塘西工业区"] ↓ 转换 |省 |市 |区 |地址 | |上海市|上海市|徐汇区|虹漕路461号58号楼5楼 | |福建省|泉州市|洛江区|万安塘西工业区 | chinese_province_city_area_mapper: built to be recognize Chinese province,city and area in simplified Chinese string, it can automaticall map area to city and map city to province. for example:: ["徐汇区虹漕路461号58号楼5楼", "泉州市洛江区万安塘西工业区"] ↓ transform |省 |市 |区 |地址 | |上海市|上海市|徐汇区|虹漕路461号58号楼5楼 | |福建省|泉州市|洛江区|万安塘西工业区 | 完整文档见该模块的Github, GitHub: `https://github.com/DQinYuan/chinese_province_city_area_mapper <https://github.com/DQinYuan/chinese_province_city_area_mapper>`_ 特点 ==== - 基于jieba分词进行匹配,同时加入了一些额外的校验匹配逻辑保证了准确率 - 因为jieba分词本身只有80%的准确率,经常会分错,所以引入了全文匹配的模式,这种模式下能够提高准确率,但会导致性能降低,关于如何开启这个模式见Github上的使用文档 - 如果地址数据比较脏的,不能指望依靠这个模块达到100%的准确,本模块只能保证尽可能地提取信息,如果想要达到100%准确率的话,最好在匹配完后再人工核验一下 - 自带完整的省,市,区三级地名及其经纬度的数据 - 支持自定义省,市,区映射 - 输出的是基于pandas的DataFrame类型的表结构,易于理解和使用 - 封装了简单的绘图功能,可以很方便地进行简单的数据可视化 - MIT 授权协议 安装说明 ======== 代码目前仅仅支持python3 pip install cpca Get Started ============ 本模块中最主要的方法是cpca.transform, 该方法可以输入任意的可迭代类型(如list,pandas的Series类型等), 然后将其转换为一个DataFrame,下面演示一个最为简单的使用方法:: location_str = ["徐汇区虹漕路461号58号楼5楼", "泉州市洛江区万安塘西工业区", "朝阳区北苑华贸城"] from cpca import * df = transform(location_str) df 输出的结果为:: 区 市 省 地址 0 徐汇区 上海市 上海市 虹漕路461号58号楼5楼 1 洛江区 泉州市 福建省 万安塘西工业区 2 朝阳区 北京市 北京市 北苑华贸城 **全文模式**: jieba分词并不能百分之百保证分词的正确性,在分词错误的情况下会造成奇怪的结果,比如下面:: location_str = ["浙江省杭州市下城区青云街40号3楼","广东省东莞市莞城区东莞大道海雅百货"] from cpca import * df = transform(location_str) df 输出的结果为:: 区 市 省 地址 城区 东莞市 广东省 莞大道海雅百货自然堂专柜 城区 杭州市 浙江省 下青云街40号3楼 这种诡异的结果因为jieba本身就将词给分错了,所以我们引入了全文模式,不进行分词,直接全文匹配,使用方法如下:: location_str = ["浙江省杭州市下城区青云街40号3楼","广东省东莞市莞城区东莞大道海雅百货"] from cpca import * df = transform(location_str, cut=False) df 输出结果:: 区 市 省 地址 下城区 杭州市 浙江省 青云街40号3楼 莞城区 东莞市 广东省 大道海雅百货 这些就完全正确了,不过全文匹配模式会造成效率低下,我默认会向前看8个字符(对应transform中的lookahead参数默认值为8),这个是比较保守的,因为有的地名会比较长(比如“新疆维吾尔族自治区”),如果你的地址库中都是些短小的省市区名的话,可以选择将lookahead设置得小一点,比如:: location_str = ["浙江省杭州市下城区青云街40号3楼","广东省东莞市莞城区东莞大道海雅百货"] from cpca import * df = transform(location_str, cut=False, lookahead=3) df 输出结果与上面一样。 如果还想知道更多的细节,请访问该 模块的github地址 `https://github.com/DQinYuan/chinese_province_city_area_mapper <https://github.com/DQinYuan/chinese_province_city_area_mapper>`_, 在那里我写了更多的细节. """ requires = ['pandas(>=0.20.0)', 'jieba(>=0.39)', ] setup(name='cpca', version='0.3.4', description='Chinese Province, City and Area Recognition Utilities', long_description=LONGDOC, author='DQinYuan', author_email='sa517067@mail.ustc.edu.cn', url='https://github.com/DQinYuan/chinese_province_city_area_mapper', license="MIT", classifiers=[ 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Natural Language :: Chinese (Simplified)', 'Programming Language :: Python :: 3.6', 'Topic :: Text Processing', 'Topic :: Text Processing :: Indexing', ], keywords='Simplified Chinese,Chinese geographic information,Chinese province city and area recognition and map', packages=['chinese_province_city_area_mapper', ''], package_dir={'chinese_province_city_area_mapper':'chinese_province_city_area_mapper', '':'.',}, #必须写成'.',而不能写成'./' install_requires = requires, )
[ "932087612@qq.com" ]
932087612@qq.com
371f634ce0aaff313e172b3b8cd67e6b98119edc
cdcaf074da4a38c2d653705225e8578e5c8a1372
/爬取天气.py
03c682424390894b35734bd1d8e09a7017f5d30c
[]
no_license
475271359/TST
f71b8632ae06f3963e9b440d8c26885fe6b9be58
d9cbd24bab66f4f92718f424a40345ae3944695c
refs/heads/master
2020-11-26T04:25:04.989105
2020-03-10T03:37:27
2020-03-10T03:37:27
228,964,176
0
0
null
null
null
null
UTF-8
Python
false
false
3,268
py
""" 目标网站: 2345天气网站 总目标:爬取北方大区所有城市的历史天气。 阶段性目标:爬取山东济南过去一年的天气数据。 输入内容: 分析网站: 待爬取页面:http://tianqi.2345.com/wea_history/54511.htm 待爬取数据:数据在js里面,http://tianqi.2345.com/t/wea_history/js/202001/54511_202001.js 构建循环,用request库,批量下载。 处理内容: 不是标准的json, 实现返回的javascript解析,得到目标数据。 对于javascript的json如何解析? 使用tqInfos = demjson.decode(data)["tqInfo"]进行解析。 输出内容: 保存为csv http://tianqi.2345.com/t/wea_history/js/202001/54511_202001.js """ #构造2019年全年的月份列表 import requests as re import demjson import numpy as np import pandas as pd global city #list = [54237, 54161, 54347, 54453, 54471, 54433, 54353, 54497, 53892, 54449, 53698, 54534, 54602, 53982, 60259, # 53986,57091, 58005, 57195, 60255, 57186, 53978, 57073, 71361, 57051, 54857, 54765, 54843, 54830, 54915, # 54827,58024,54945, 54938, 54828, 54823, 54714] list = [54347,54161] def getHTMLText(url): # 定义了一个函数,用于获取html的文本。 try: r = re.get(url, timeout=30) r.raise_for_status() # 如果状态不是200,引发HTTPError异常。 r.encoding = r.apparent_encoding # 从内容中分析,修正代码的编码方式。 return r.text except: return "产生异常" for x in list: year = 2019 ymd = [] bWendu = [] yWendu = [] tianqi = [] fengxiang = [] fengli = [] aqi = [] aqiInfo = [] aqiLevel = [] city_list = [] all_datas = [] datas = [] months = ["{:d}{:0>2d}".format(year, month + 1) for month in range(12)] # 列表生成式 urls = ["http://tianqi.2345.com/t/wea_history/js/{}/".format(month) +str(x)+"_{}.js".format(month) for month in months] # 列表生成器 for url in urls: data = getHTMLText(url).lstrip("var weather_str=").rsplit(";") datas.append(data[0]) for data in datas: tqInfos = demjson.decode(data)["tqInfo"] city = demjson.decode(data)["city"] all_datas.extend(x for x in tqInfos if len(x)>0) for y in range(len(all_datas)): ymd.append(all_datas[y].get('ymd')) bWendu.append(all_datas[y].get('bWendu')) yWendu.append(all_datas[y].get('yWendu')) tianqi.append(all_datas[y].get('tianqi')) fengxiang.append(all_datas[y].get('fengxiang')) fengli.append(all_datas[y].get('fengli')) aqi.append(all_datas[y].get('aqi')) aqiInfo.append(all_datas[y].get('aqiInfo')) aqiLevel.append(all_datas[y].get('aqiLevel')) city_list.append(city) Tianqi_np=np.array([ymd,bWendu,yWendu,tianqi,fengxiang,fengli,aqi,aqiInfo,aqiLevel,city_list]) Tianqi_df = pd.DataFrame(Tianqi_np,index=["ymd","bWendu","yWendu","tianqi","fengxiang","fengli","aqi","aqiInfo","aqiLevel","city_list"]) Tianqi_df=pd.DataFrame(Tianqi_df.values.T, index=Tianqi_df.columns, columns=Tianqi_df.index) Tianqi_df.to_excel("C:/Users/86132/Desktop/"+city+".xlsx")
[ "noreply@github.com" ]
noreply@github.com
f5a8345f6e04ba694a886993167e2ae60ef1dcd4
8a6a96d487a0933293a563b04c24d08907608920
/app/auth.py
85c5eecc234752549804a0b229487c55dc6041ed
[]
no_license
shuxue051/ninja
21ec07318f8c868e61659b977edce1ebdefc8e97
ccb9724ff6329b82e18f839bb76246cae899fe99
refs/heads/master
2021-01-10T21:04:20.004674
2014-09-24T02:50:25
2014-09-24T02:50:25
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,941
py
import os, datetime from flask import current_app, Blueprint, render_template, abort, request, flash, redirect, url_for from jinja2 import TemplateNotFound from app import login_manager, flask_bcrypt from flask.ext.login import (current_user, login_required, login_user, logout_user, confirm_login, fresh_login_required) import forms from User import User auth_flask_login = Blueprint('auth_flask_login', __name__, template_folder='templates') @auth_flask_login.route("/login", methods=["GET", "POST"]) def login(): if request.method == "POST" and "email" in request.form: email = request.form["email"] userObj = User() user = userObj.get_by_email_w_password(email) if user and flask_bcrypt.check_password_hash(user.password,request.form["password"]) and user.is_active(): remember = request.form.get("remember", "no") == "yes" if login_user(user, remember=remember): flash("Logged in!") return redirect('/notes/create') else: flash("unable to log you in") return render_template("/auth/login.html") # # Route disabled - enable route to allow user registration. # @auth_flask_login.route("/register", methods=["GET","POST"]) def register(): registerForm = forms.SignupForm(request.form) current_app.logger.info(request.form) if request.method == 'POST' and registerForm.validate() == False: current_app.logger.info(registerForm.errors) return "uhoh registration error" elif request.method == 'POST' and registerForm.validate(): email = request.form['email'] # generate password hash password_hash = flask_bcrypt.generate_password_hash(request.form['password']) # prepare User user = User(email,password_hash) print user try: user.save() if login_user(user, remember="no"): flash("Logged in!") return redirect('/') else: flash("unable to log you in") except: flash("unable to register with that email address") current_app.logger.error("Error on registration - possible duplicate emails") # prepare registration form # registerForm = RegisterForm(csrf_enabled=True) templateData = { 'form' : registerForm } return render_template("/auth/register.html", **templateData) @auth_flask_login.route("/reauth", methods=["GET", "POST"]) @login_required def reauth(): if request.method == "POST": confirm_login() flash(u"Reauthenticated.") return redirect(request.args.get("next") or '/admin') templateData = {} return render_template("/auth/reauth.html", **templateData) @auth_flask_login.route("/logout") @login_required def logout(): logout_user() flash("Logged out.") return redirect('/login') @login_manager.unauthorized_handler def unauthorized_callback(): return redirect('/login') @login_manager.user_loader def load_user(id): if id is None: redirect('/login') user = User() user.get_by_id(id) if user.is_active(): return user else: return None
[ "shuxue051@gmail.com" ]
shuxue051@gmail.com
4816b6ce56b6ba10760fc6ec50b511666a0ef942
c0f5d309576f791f8cc062e2d0cad340eec41d7d
/3.py
846552142673f67774ae9cc5803b41248ec09248
[]
no_license
mjjin1214/algorithm
fa91455ab792c38d01fd210c12e53e50f516eb55
423119406061443939b4b966c7d9f1513544dd03
refs/heads/master
2020-04-22T19:31:23.981387
2019-04-05T07:58:10
2019-04-05T07:58:10
170,610,108
0
0
null
null
null
null
UTF-8
Python
false
false
456
py
import sys sys.stdin = open('input2.txt') def subset(n, su): global visit, count if n == len(score): if not visit & (1<<su): visit ^= (1<<su) count += 1 return subset(n+1, su+score[n]) subset(n+1, su) T = int(input()) for t in range(T): N = int(input()) score = list(set(map(int, input().split()))) visit = count = 0 subset(0, 0) print('#{} {}'.format(t+1, count+N-len(score)))
[ "moc0etan@gmail.com" ]
moc0etan@gmail.com
a19dea9f4e498c18ae262efb60f03b1b275e63ba
f43df719ae5cc238c450f01959e4cb3625ae1c82
/config_make.py
d96d27bb8e73dd8fb5491ba43c7d0623f1a1fee9
[]
no_license
kenshow8/eth_erc20_token_python_api
8f10e5aaa503730b4abc66ebd8f1779b97e021f0
4390d47bcb574205230fefb8cb9b2e4b238ac27e
refs/heads/master
2022-09-29T14:09:10.058167
2020-05-31T08:08:14
2020-05-31T08:08:14
268,100,598
0
0
null
null
null
null
UTF-8
Python
false
false
796
py
# coding:utf-8 import os import configparser from pathlib import Path file_dir = Path(__file__).parent file_name = 'config/setting.ini' file_path = file_dir.joinpath(file_name) config = configparser.ConfigParser() # コントラクト設定 config_web3 = 'web3_config' config.add_section(config_web3) ## web3 設定 ## config.set(config_web3, 'web3_url', 'http://127.0.0.1:7545') # contractのデプロイ先に応じて変更してください。 config.set(config_web3, 'contract_address', '0x238f011262D73a07c0bfACdf5f851CE467bc94ee') # contract addressを入力してください。 config.set(config_web3, 'owner_address', '0xB66d64EF0fACCebFd6F5E10Ece3dcBd3a65B82F1') # contractのowner addressを入力してください。 with open(file_path, 'w') as file: config.write(file)
[ "wmapp.kj@gmail.com" ]
wmapp.kj@gmail.com
2f4d530d9958802e44c5a71b92bc8a6bf4fd5a84
0149c07b7c1bdd2d56aecf9934e16ec9c5b163c7
/oc_ocdm/test/storer/test_storer.py
5be49fe468ed9e80b576ea191f16708cec2154aa
[ "ISC" ]
permissive
opencitations/oc_ocdm
fac6903403b4149cafd080a88d761ebb44fb91cf
26298e81fd4903ad60edb1d824f2c7abbdc88656
refs/heads/master
2023-07-07T23:53:38.396919
2023-06-22T15:20:16
2023-06-22T15:20:16
322,327,342
2
6
ISC
2023-03-14T08:41:54
2020-12-17T14:56:50
Python
UTF-8
Python
false
false
11,370
py
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright (c) 2016, Silvio Peroni <essepuntato@gmail.com> # # Permission to use, copy, modify, and/or distribute this software for any purpose # with or without fee is hereby granted, provided that the above copyright notice # and this permission notice appear in all copies. # # THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES WITH # REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF MERCHANTABILITY AND # FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY SPECIAL, DIRECT, INDIRECT, # OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES WHATSOEVER RESULTING FROM LOSS OF USE, # DATA OR PROFITS, WHETHER IN AN ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS # ACTION, ARISING OUT OF OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS # SOFTWARE. import json import os import unittest from platform import system from shutil import rmtree from zipfile import ZipFile from rdflib import ConjunctiveGraph, URIRef, compare from oc_ocdm.graph.graph_set import GraphSet from oc_ocdm.prov.prov_set import ProvSet from oc_ocdm.storer import Storer class TestStorer(unittest.TestCase): def setUp(self): self.resp_agent = "http://resp_agent.test/" self.base_iri = "http://test/" self.graph_set = GraphSet(self.base_iri, "", "060", False) self.prov_set = ProvSet(self.graph_set, self.base_iri, "", False) self.br = self.graph_set.add_br(self.resp_agent) def tearDown(self): rmtree(os.path.join("oc_ocdm", "test", "storer", "data")) def test_store_graphs_in_file(self): base_dir = os.path.join("oc_ocdm", "test", "storer", "data", "rdf") + os.sep is_unix = system() != "Windows" with self.subTest("output_format=json-ld, zip_output=True"): modified_entities = self.prov_set.generate_provenance() prov_storer = Storer(self.prov_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='json-ld', zip_output=True) storer = Storer(self.graph_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='json-ld', zip_output=True, modified_entities=modified_entities) storer.store_all(base_dir, self.base_iri) prov_storer.store_all(base_dir, self.base_iri) self.graph_set.commit_changes() with ZipFile(os.path.join(base_dir, "br", "060", "10000", "1000.zip"), mode="r") as archive: with archive.open("1000.json") as f: data = json.load(f) self.assertEqual(data, [{'@graph': [{'@id': 'http://test/br/0601', '@type': ['http://purl.org/spar/fabio/Expression']}], '@id': 'http://test/br/'}]) with ZipFile(os.path.join(base_dir, "br", "060", "10000", "1000", "prov", "se.zip"), mode="r") as archive: with archive.open("se.json") as f: data = [{g:[{k:v for k,v in datum.items() if k != "http://www.w3.org/ns/prov#generatedAtTime"} for datum in data] if g == "@graph" else data for g, data in graph.items()} for graph in json.load(f)] self.assertEqual(data, [{'@graph': [{ '@id': 'http://test/br/0601/prov/se/1', '@type': ['http://www.w3.org/ns/prov#Entity'], 'http://purl.org/dc/terms/description': [{'@value': "The entity 'http://test/br/0601' has been created."}], 'http://www.w3.org/ns/prov#specializationOf': [{'@id': 'http://test/br/0601'}], 'http://www.w3.org/ns/prov#wasAttributedTo': [{'@id': 'http://resp_agent.test/'}]}], '@id': 'http://test/br/0601/prov/'}]) if is_unix: self.assertTrue(os.path.exists(os.path.join(base_dir, "br", "060", "10000", "1000.zip.lock"))) self.assertTrue(os.path.exists(os.path.join(base_dir, "br", "060", "10000", "1000", "prov", "se.zip.lock"))) with self.subTest("output_format=json-ld, zip_output=False"): base_dir_1 = os.path.join("oc_ocdm", "test", "storer", "data", "rdf_1") + os.sep storer = Storer(self.graph_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='json-ld', zip_output=False) self.prov_set.generate_provenance() prov_storer = Storer(self.prov_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='json-ld', zip_output=False) storer.store_all(base_dir_1, self.base_iri) prov_storer.store_all(base_dir_1, self.base_iri) self.graph_set.commit_changes() with open(os.path.join(base_dir_1, "br", "060", "10000", "1000.json")) as f: data = json.load(f) self.assertEqual(data, [{'@graph': [{'@id': 'http://test/br/0601', '@type': ['http://purl.org/spar/fabio/Expression']}], '@id': 'http://test/br/'}]) with open(os.path.join(base_dir_1, "br", "060", "10000", "1000", "prov", "se.json")) as f: data = [{g:[{k:v for k,v in datum.items() if k != "http://www.w3.org/ns/prov#generatedAtTime"} for datum in data] if g == "@graph" else data for g, data in graph.items()} for graph in json.load(f)] self.assertEqual(data, [{'@graph': [{ '@id': 'http://test/br/0601/prov/se/1', '@type': ['http://www.w3.org/ns/prov#Entity'], 'http://purl.org/dc/terms/description': [{'@value': "The entity 'http://test/br/0601' has been created."}], 'http://www.w3.org/ns/prov#specializationOf': [{'@id': 'http://test/br/0601'}], 'http://www.w3.org/ns/prov#wasAttributedTo': [{'@id': 'http://resp_agent.test/'}]}], '@id': 'http://test/br/0601/prov/'}]) if is_unix: self.assertTrue(os.path.exists(os.path.join(base_dir_1, "br", "060", "10000", "1000.json.lock"))) self.assertTrue(os.path.exists(os.path.join(base_dir_1, "br", "060", "10000", "1000", "prov", "se.json.lock"))) with self.subTest("output_format=nquads, zip_output=True"): base_dir_2 = os.path.join("oc_ocdm", "test", "storer", "data", "rdf_2") + os.sep storer = Storer(self.graph_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='nquads', zip_output=True) self.prov_set.generate_provenance() prov_storer = Storer(self.prov_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='nquads', zip_output=True) storer.store_all(base_dir_2, self.base_iri) prov_storer.store_all(base_dir_2, self.base_iri) self.graph_set.commit_changes() with ZipFile(os.path.join(base_dir_2, "br", "060", "10000", "1000.zip"), mode="r") as archive: with archive.open("1000.nt") as f: data = f.read().decode("utf-8") self.assertEqual(data, "<http://test/br/0601> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://purl.org/spar/fabio/Expression> <http://test/br/> .\n\n") with ZipFile(os.path.join(base_dir_2, "br", "060", "10000", "1000", "prov", "se.zip"), mode="r") as archive: with archive.open("se.nq") as f: data = f.read().decode("utf-8") data_g = ConjunctiveGraph() expected_data_g = ConjunctiveGraph() data_g.parse(data=data, format="nquads") expected_data_g.parse(data=""" <http://test/br/0601/prov/se/1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://www.w3.org/ns/prov#Entity> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://www.w3.org/ns/prov#specializationOf> <http://test/br/0601> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://www.w3.org/ns/prov#wasAttributedTo> <http://resp_agent.test/> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://purl.org/dc/terms/description> "The entity 'http://test/br/0601' has been created." <http://test/br/0601/prov/> . """, format="nquads") for s, p, o, c in data_g.quads(): if p == URIRef("http://www.w3.org/ns/prov#generatedAtTime"): data_g.remove((s, p, o, c)) self.assertTrue(compare.isomorphic(data_g, expected_data_g)) if is_unix: self.assertTrue(os.path.exists(os.path.join(base_dir_2, "br", "060", "10000", "1000.zip.lock"))) self.assertTrue(os.path.exists(os.path.join(base_dir_2, "br", "060", "10000", "1000", "prov", "se.zip.lock"))) with self.subTest("output_format=nquads, zip_output=False"): base_dir_3 = os.path.join("oc_ocdm", "test", "storer", "data", "rdf_3") + os.sep storer = Storer(self.graph_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='nquads', zip_output=False) self.prov_set.generate_provenance() prov_storer = Storer(self.prov_set, context_map={}, dir_split=10000, n_file_item=1000, default_dir="_", output_format='nquads', zip_output=False) storer.store_all(base_dir_3, self.base_iri) prov_storer.store_all(base_dir_3, self.base_iri) self.graph_set.commit_changes() prov_unzipped = ConjunctiveGraph() expected_prov_unzipped = ConjunctiveGraph() with open(os.path.join(base_dir_3, "br", "060", "10000", "1000.nt"), "r", encoding="utf-8") as f: data_unzipped = f.read() prov_unzipped.parse(source=os.path.join(base_dir_3, "br", "060", "10000", "1000", "prov", "se.nq"), format="nquads") expected_prov_unzipped.parse(data=""" <http://test/br/0601/prov/se/1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://www.w3.org/ns/prov#Entity> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://www.w3.org/ns/prov#specializationOf> <http://test/br/0601> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://www.w3.org/ns/prov#wasAttributedTo> <http://resp_agent.test/> <http://test/br/0601/prov/> . <http://test/br/0601/prov/se/1> <http://purl.org/dc/terms/description> "The entity 'http://test/br/0601' has been created." <http://test/br/0601/prov/> . """, format="nquads") for s, p, o, c in prov_unzipped.quads(): if p == URIRef("http://www.w3.org/ns/prov#generatedAtTime"): prov_unzipped.remove((s, p, o, c)) self.assertEqual(data_unzipped, "<http://test/br/0601> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://purl.org/spar/fabio/Expression> <http://test/br/> .\n\n") self.assertTrue(compare.isomorphic(prov_unzipped, expected_prov_unzipped)) if is_unix: self.assertTrue(os.path.exists(os.path.join(base_dir_3, "br", "060", "10000", "1000.nt.lock"))) self.assertTrue(os.path.exists(os.path.join(base_dir_3, "br", "060", "10000", "1000", "prov", "se.nq.lock"))) if __name__ == '__main__': unittest.main()
[ "arcangelomas@gmail.com" ]
arcangelomas@gmail.com
587c0100e02d4abc95b9fb79c169f06d3ed8b47b
891c267df1228d42153ff1a36d6b92b58a32f435
/schemes.py
ad91237c7020256b802593dc48fb3b750e821042
[]
no_license
ellawang44/numerical-advection-eq-MATH2306-
eaa6be239e1d27f1aed2cb29bba4fdfde305a081
de410631ec9d2fcb68c7aa2405f2cd785cb1c604
refs/heads/master
2021-07-21T12:17:50.392999
2017-10-25T02:15:15
2017-10-25T02:15:15
106,899,828
0
0
null
null
null
null
UTF-8
Python
false
false
568
py
def ftbs(jn1, j, C_r): """Forward Time Backwards Space Scheme""" return j - C_r*(j - jn1) def ftcs(jn1, j, j1, C_r): """Forward Time Centered Space Scheme""" return j - C_r/2*(j1 - jn1) def ftfs(j, j1, C_r): """Forwards Time Forwards Space Scheme""" return j - C_r*(j1 - j) def lax_wendroff(jn1, j, j1, C_r): """Lax Wendroff Scheme""" return j - C_r/2*(j1 - jn1) + C_r**2/2*(j1 - 2*j + jn1) def beam_warming(jn2, jn1, j, C_r): """Beam Warming Scheme""" return j - C_r/2*(3*j - 4*jn1 + jn2) + C_r**2/2*(j - 2*jn1 + jn2)
[ "u6083620@anu.edu.au" ]
u6083620@anu.edu.au
ae9c47cbad8ea97eded1a2947f4a8a01472b1951
be2fe3d75f5d08d44d207de42309aefac90f8c64
/O Big.py
02c8fc7d2f7e6ff0e6cb59e6ce33be3bbad2de42
[]
no_license
lucasilvabc/CS241
be18016862b420da0a7d44d9d6ab49aa76111e81
e94487a7d4b1e4b282de8df419fa2abad61f2087
refs/heads/main
2023-04-06T20:11:01.283402
2021-04-19T23:42:58
2021-04-19T23:42:58
344,626,769
0
0
null
null
null
null
UTF-8
Python
false
false
178
py
def contains(haystack, needle): # Does the haystack contain the needle? for item in haystack: if item == needle: return True return False
[ "noreply@github.com" ]
noreply@github.com
f7bb5b008461cd4f51770163a3cf7e600d784405
81c5c07e1144747dc0e98f8dffb287a69be1eba7
/score_mcc_bin.py
686c4e86fcab42e4f12a69f6f893e59e1cfe31ee
[]
no_license
twistedmove/e2e_antispoofing
acbb9ec5bc4454c1698fc355d0c0fee3bf70006e
686dfb515b2c568a1006136f56bbaad0419f0787
refs/heads/master
2020-09-07T10:41:12.024794
2019-10-06T19:28:19
2019-10-06T19:28:19
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,454
py
import argparse import numpy as np import glob import torch import torch.nn.functional as F import os from kaldi_io import read_mat_scp import model as model_ import scipy.io as sio from utils import compute_eer_labels, set_device, read_trials, get_freer_gpu def prep_feats(data_): #data_ = ( data_ - data_.mean(0) ) / data_.std(0) features = data_.T if features.shape[1]<50: mul = int(np.ceil(50/features.shape[1])) features = np.tile(features, (1, mul)) features = features[:, :50] return torch.from_numpy(features[np.newaxis, np.newaxis, :, :]).float() if __name__ == '__main__': parser = argparse.ArgumentParser(description='Compute scores for mcc model') parser.add_argument('--path-to-data', type=str, default='./data/feats.scp', metavar='Path', help='Path to input data') parser.add_argument('--trials-path', type=str, default='./data/trials', metavar='Path', help='Path to trials file') parser.add_argument('--cp-path', type=str, default=None, metavar='Path', help='Path for file containing model') parser.add_argument('--out-path', type=str, default='./out.txt', metavar='Path', help='Path to output hdf file') parser.add_argument('--model', choices=['lstm', 'resnet', 'resnet_pca', 'lcnn_9', 'lcnn_29', 'lcnn_9_pca', 'lcnn_29_pca', 'lcnn_9_prodspec', 'lcnn_9_icqspec', 'lcnn_9_CC', 'lcnn_29_CC', 'resnet_CC'], default='lcnn_9', help='Model arch') parser.add_argument('--n-classes', type=int, default=-1, metavar='N', help='Number of classes for the mcc case (default: binary classification)') parser.add_argument('--no-cuda', action='store_true', default=False, help='Disables GPU use') parser.add_argument('--no-output-file', action='store_true', default=False, help='Disables writing scores into out file') parser.add_argument('--no-eer', action='store_true', default=False, help='Disables computation of EER') parser.add_argument('--eval', action='store_true', default=False, help='Enables eval trials reading') parser.add_argument('--ncoef', type=int, default=90, metavar='N', help='Number of cepstral coefs (default: 90)') parser.add_argument('--init-coef', type=int, default=0, metavar='N', help='First cepstral coefs (default: 0)') args = parser.parse_args() args.cuda = True if not args.no_cuda and torch.cuda.is_available() else False if args.cp_path is None: raise ValueError('There is no checkpoint/model path. Use arg --cp-path to indicate the path!') if os.path.isfile(args.out_path): os.remove(args.out_path) print(args.out_path + ' Removed') print('Cuda Mode is: {}'.format(args.cuda)) print('Selected model is: {}'.format(args.model)) if args.cuda: device = get_freer_gpu() if args.model == 'lstm': model = model_.cnn_lstm(nclasses=args.n_classes) elif args.model == 'resnet': model = model_.ResNet(nclasses=args.n_classes) elif args.model == 'resnet_pca': model = model_.ResNet_pca(nclasses=args.n_classes) elif args.model == 'lcnn_9': model = model_.lcnn_9layers(nclasses=args.n_classes) elif args.model == 'lcnn_29': model = model_.lcnn_29layers_v2(nclasses=args.n_classes) elif args.model == 'lcnn_9_pca': model = model_.lcnn_9layers_pca(nclasses=args.n_classes) elif args.model == 'lcnn_29_pca': model = model_.lcnn_29layers_v2_pca(nclasses=args.n_classes) elif args.model == 'lcnn_9_icqspec': model = model_.lcnn_9layers_icqspec(nclasses=args.n_classes) elif args.model == 'lcnn_9_prodspec': model = model_.lcnn_9layers_prodspec(nclasses=args.n_classes) elif args.model == 'lcnn_9_CC': model = model_.lcnn_9layers_CC(nclasses=args.n_classes, ncoef=args.ncoef, init_coef=args.init_coef) elif args.model == 'lcnn_29_CC': model = model_.lcnn_29layers_CC(nclasses=args.n_classes, ncoef=args.ncoef, init_coef=args.init_coef) elif args.model == 'resnet_CC': model = model_.ResNet_CC(nclasses=args.n_classes, ncoef=args.ncoef, init_coef=args.init_coef) print('Loading model') ckpt = torch.load(args.cp_path, map_location = lambda storage, loc: storage) model.load_state_dict(ckpt['model_state'], strict=False) model.eval() print('Model loaded') print('Loading data') if args.eval: test_utts = read_trials(args.trials_path, eval_=args.eval) else: test_utts, attack_type_list, label_list = read_trials(args.trials_path, eval_=args.eval) data = { k:m for k,m in read_mat_scp(args.path_to_data) } print('Data loaded') print('Start of scores computation') score_list = [] with torch.no_grad(): for i, utt in enumerate(test_utts): print('Computing score for utterance '+ utt) feats = prep_feats(data[utt]) try: if args.cuda: feats = feats.to(device) model = model.to(device) score = 1.-F.softmax(model.forward(feats), dim=1)[:,1:].sum().item() except: feats = feats.cpu() model = model.cpu() score = 1.-F.softmax(model.forward(feats), dim=1)[:,1:].sum().item() score_list.append(score) print('Score: {}'.format(score_list[-1])) if not args.no_output_file: print('Storing scores in output file:') print(args.out_path) with open(args.out_path, 'w') as f: if args.eval: for i, utt in enumerate(test_utts): f.write("%s" % ' '.join([utt, str(score_list[i])+'\n'])) else: for i, utt in enumerate(test_utts): f.write("%s" % ' '.join([utt, attack_type_list[i], label_list[i], str(score_list[i])+'\n'])) if not args.no_eer and not args.eval: print('EER: {}'.format(compute_eer_labels(label_list, score_list))) print('All done!!')
[ "joaomonteirof@gmail.com" ]
joaomonteirof@gmail.com
6c1a65c48e423d27b57f75c12495f1387113b2df
3a4ecc7f7db23e92b7f8e406d3a73a77c53e1472
/src/chapter02/c2_df1.py
6026dc4be928099e2aae64f9481d1a1313b6da19
[]
no_license
Parthi3610/pandas_workbook
b89b768f2d99cb5f042accbaebc1470bc350565e
6e61f67db42ff708b6f40a9fb96b755085fe6587
refs/heads/main
2023-01-19T08:52:37.000264
2020-11-06T03:58:38
2020-11-06T03:58:38
303,878,391
0
0
null
null
null
null
UTF-8
Python
false
false
2,071
py
import pandas as pd import numpy as np movies = pd.read_csv("C:/Users/preet/PycharmProjects/pandas_workbook/src/data/movie.csv") movie_act_dir = movies[ [ "actor_1_name", "actor_2_name", "actor_3_name", "director_name" ] ] print(movie_act_dir.head()) print(type(movies[["director_name"]])) print(type(movies["director_name"])) print(type(movies.loc[:,["director_name"]])) print(type(movies.loc[:,"director_name"])) def shorten(col): return ( str(col).replace("facebook_likes","fb").replace("_for_reviews","") ) movies = movies.rename(columns = shorten) print (movies.dtypes.value_counts()) print(movies.select_dtypes(include = "int64").head()) print(movies.select_dtypes(include = "float64").head()) print(movies.select_dtypes(include = "number").head()) print(movies.select_dtypes(include = ["int64","object"]).head()) print(movies.filter(like = 'fb').head()) print(movies.columns) cat_core = [ "movie_title", "title_year", "content_rating", "genres", ] cat_people = [ "director_name", "actor_1_name", "actor_2_name", "actor_3_name", ] cat_other = [ "color", "country", "language", "plot_keywords", "movie_imdb_link", ] cont_fb = [ "director_fb", "actor_1_fb", "actor_2_fb", "actor_3_fb", "cast_total_fb", "movie_fb", ] cont_finance = ["budget", "gross"] cont_num_reviews = [ "num_voted_users", "num_user", "num_critic", ] cont_other = [ "imdb_score", "duration", "aspect_ratio", "facenumber_in_poster", ] new_col_order = (cat_core+cat_people+cat_other+cont_fb+cont_finance+cont_num_reviews+cont_other) print(set(movies.columns) == set(new_col_order)) print(movies[new_col_order].head()) print(movies.describe()) print(movies.describe().T) print(movies.describe(percentiles=[0.01,0.3,1]).T) print(movies.min(skipna=False)) print(movies.isnull().head()) print(movies.isnull().sum().head()) print(movies.isnull().sum().sum()) print(movies.select_dtypes(["object"]).fillna("").max())
[ "parthi_3610@yahoo.co.in" ]
parthi_3610@yahoo.co.in
ed321ee3351dd686798f469165a774a3b522d685
8541f96e5f48780050a7995ed735c70d4f3c2642
/test/test_validation.py
5fb04f8372aea80988cd6820c8839dc4e3d5c76a
[ "LicenseRef-scancode-public-domain", "CC0-1.0" ]
permissive
influence-usa/lobbying_federal_domestic
ede64752f81b6a8496a9f6b31b3f2a6c3a5f485e
117a1991e01a203c2c181c1c6163b044c4d5c56f
refs/heads/master
2021-01-02T23:02:57.140303
2016-11-30T22:47:46
2016-11-30T22:47:46
17,500,504
7
2
null
2014-05-08T21:32:29
2014-03-07T02:39:15
Python
UTF-8
Python
false
false
3,903
py
import json import pytest import validictory from utils.validate import validate_uuid, validate_url, validate_email @pytest.fixture def uuid_fixture(): data = json.loads(''' { "uuidInt": 117574695023396164616661330147169357159, "uuidHex": "054a4828074e45f293a3a7ffbcd43bfb", "uuidCanon": "054a4828-074e-45f2-93a3-a7ffbcd43bfb" }''') schema = { "title": "My test schema", "properties": { "uuidHex": { "format": "uuid_hex" }, "uuidInt": { "format": "uuid_int" }, "uuidCanon": { "format": "uuid_hex" } } } return {'data': data, 'schema': schema} def test_validate_uuid(uuid_fixture): uuid_data = uuid_fixture['data'] uuid_schema = uuid_fixture['schema'] formatdict = {"uuid_hex": validate_uuid, "uuid_int": validate_uuid} # Make sure good data validates validictory.validate(uuid_data, uuid_schema, format_validators=formatdict) # Make sure bad data doesn't with pytest.raises(validictory.ValidationError): bad_data = uuid_data.copy() bad_data['uuidHex'] = 'not_a_uuid' validictory.validate(bad_data, uuid_schema, format_validators=formatdict) @pytest.fixture def url_fixture(): data = json.loads(''' { "test_http": "http://sunlightfoundation.com/api", "test_https": "https://www.aal-usa.com", "test_ftp": "ftp://ftp.fec.gov/FEC/" }''') schema = { "title": "Url test schema", "properties": { "test_http": { "format": "url_http" }, "test_https": { "format": "url_http" }, "test_ftp": { "format": "url_ftp" }, } } return {'data': data, 'schema': schema} def test_validate_url(url_fixture): url_data = url_fixture['data'] url_schema = url_fixture['schema'] formatdict = {"url_http": validate_url, "url_ftp": validate_url} # Make sure good data validates validictory.validate(url_data, url_schema, format_validators=formatdict) # Make sure bad data doesn't bad_egs = zip(['test_http', 'test_https', 'test_ftp'], ['sunlightfoundation.com', 'https:/www.aal-usa.com', 'ftp:://ftp.fec.fgov/FEC/']) print bad_egs for field, bad_eg in bad_egs: with pytest.raises(validictory.ValidationError): bad_data = url_data.copy() bad_data[field] = bad_eg print bad_eg validictory.validate(bad_data, url_schema, format_validators=formatdict) @pytest.fixture def email_fixture(): data = json.loads(''' { "email": "blannon@sunlightfoundation.com" }''') schema = { "title": "Email test schema", "properties": { "email": { "format": "email" } } } return {'data': data, 'schema': schema} def test_validate_email(email_fixture): email_data = email_fixture['data'] email_schema = email_fixture['schema'] formatdict = {"email": validate_email} # Make sure good data validates validictory.validate(email_data, email_schema, format_validators=formatdict) # Make sure bad data doesn't with pytest.raises(validictory.FieldValidationError): bad_data = email_data.copy() bad_data['email'] = 'bobby bear at gmail.com' validictory.validate(bad_data, email_schema, format_validators=formatdict)
[ "blannon@gmail.com" ]
blannon@gmail.com
9872a4700e6e9c1bd2a6de233fa4c578321b2559
ac6fdcfe85b72bcf1b719bdea3ce7f9954217fd3
/zap_transformers.py
99dcb9b97c33bf38a861744f3fe4f97081fc6b5d
[]
no_license
giuice/ds-zap-challenge
084510963ee413b8d62aafcb600cfebd1b86d6e6
34eb56c1ee74d9af02a668522bb97fea1ac95c9e
refs/heads/master
2020-06-25T23:08:32.085580
2019-08-16T21:07:48
2019-08-16T21:07:48
199,449,612
0
0
null
null
null
null
UTF-8
Python
false
false
21,621
py
# Reference # import numpy as np import pandas as pd import warnings from sklearn.base import BaseEstimator, TransformerMixin from sklearn.utils.validation import check_is_fitted from base_transformers import BaseCategoricalEncoder,BaseCategoricalTransformer, _define_variables, BaseNumericalTransformer from sklearn.feature_extraction import DictVectorizer from sklearn.preprocessing import FunctionTransformer, StandardScaler, RobustScaler from sklearn.preprocessing import Imputer, MultiLabelBinarizer class DFStandardScaler(TransformerMixin): # StandardScaler but for pandas DataFrames def __init__(self): self.ss = None self.mean_ = None self.scale_ = None def fit(self, X, y=None): self.ss = StandardScaler() self.ss.fit(X) self.mean_ = pd.Series(self.ss.mean_, index=X.columns) self.scale_ = pd.Series(self.ss.scale_, index=X.columns) return self def transform(self, X): # assumes X is a DataFrame Xss = self.ss.transform(X) Xscaled = pd.DataFrame(Xss, index=X.index, columns=X.columns) return Xscaled class DFRobustScaler(BaseNumericalTransformer): # RobustScaler but for pandas DataFrames def __init__(self, variables=None): self.variables = _define_variables(variables) self.rs = None self.center_ = None self.scale_ = None def fit(self, X, y=None): self.rs = RobustScaler() var = self.variables self.rs.fit(X[var]) self.center_ = pd.Series(self.rs.center_, index=X[var].columns) self.scale_ = pd.Series(self.rs.scale_, index=X[var].columns) return self def transform(self, X): # assumes X is a DataFrame #for feature in self.variables: var = self.variables X[var] = self.rs.transform(X[var]) #Xscaled = pd.DataFrame(Xrs, index=X.index, columns=X.columns) return X class ColumnExtractor(TransformerMixin): def __init__(self, cols): self.cols = cols def fit(self, X, y=None): # stateless transformer return self def transform(self, X): # assumes X is a DataFrame Xcols = X[self.cols] return Xcols class ZeroFillTransformer(TransformerMixin): def fit(self, X, y=None): # stateless transformer return self def transform(self, X): # assumes X is a DataFrame Xz = X.fillna(value=0) return Xz class CountFrequencyCategoricalEncoder(BaseCategoricalEncoder): """ The CountFrequencyCategoricalEncoder() replaces categories by the count of observations per category or by the percentage of observations per category. For example in the variable colour, if 10 observations are blue, blue will be replaced by 10. Alternatively, if 10% of the observations are blue, blue will be replaced by 0.1. The CountFrequencyCategoricalEncoder() will encode only categorical variables (type 'object'). A list of variables can be passed as an argument. If no variables are passed as argument, the encoder will only encode categorical variables (object type) and ignore the rest. The encoder first maps the categories to the numbers for each variable (fit). The encoder then transforms the categories to those mapped numbers (transform). Parameters ---------- encoding_method : str, default='count' Desired method of encoding. 'count': number of observations per category 'frequency' : percentage of observations per category variables : list The list of categorical variables that will be encoded. If None, the encoder will find and transform all object type variables. Attributes ---------- encoder_dict_: dictionary The dictionary containing the {count / frequency: category} pairs used to replace categories for every variable. """ def __init__(self, encoding_method = 'count', variables = None): if encoding_method not in ['count', 'frequency']: raise ValueError("encoding_method takes only values 'count' and 'frequency'") self.encoding_method = encoding_method self.variables = _define_variables(variables) def fit(self, X, y = None): """ Learns the numbers that should be used to replace the categories in each variable. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The training input samples. Can be the entire dataframe, not just seleted variables. y : None y is not needed in this encoder, yet the sklearn pipeline API requires this parameter for checking. You can either leave it as None or pass y. """ # brings the variables from the BaseEncoder super().fit(X, y) self.encoder_dict_ = {} for var in self.variables: if self.encoding_method == 'count': self.encoder_dict_[var] = X[var].value_counts().to_dict() elif self.encoding_method == 'frequency': n_obs = np.float(len(X)) self.encoder_dict_[var] = (X[var].value_counts() / n_obs).to_dict() if len(self.encoder_dict_)==0: raise ValueError('Encoder could not be fitted. Check that correct parameters and dataframe were passed during training') self.input_shape_ = X.shape return self class OrdinalCategoricalEncoder(BaseCategoricalEncoder): """ The OrdinalCategoricalEncoder() replaces categories by ordinal numbers (0, 1, 2, 3, etc). The numbers can be ordered based on the mean of the target per category, or assigned arbitrarily. For the ordered ordinal encoding for example in the variable colour, if the mean of the target for blue, red and grey is 0.5, 0.8 and 0.1 respectively, blue is replaced by 1, red by 2 and grey by 0. For the arbitrary ordinal encoding the numbers will be assigned arbitrarily to the categories, on a first seen first served basis. The Encoder will encode only categorical variables (type 'object'). A list of variables can be passed as an argument. If no variables are passed as argument, the encoder will only encode categorical variables (object type) and ignore the rest. The encoder first maps the categories to the numbers for each variable (fit). The encoder then transforms the categories to the mapped numbers (transform). Parameters ---------- encoding_method : str, default='ordered' Desired method of encoding. 'ordered': the categories are numbered in ascending order according to the target mean per category. 'arbitrary' : categories are numbered arbitrarily. variables : list, default=None The list of categorical variables that will be encoded. If None, the encoder will find and select all object type variables. Attributes ---------- encoder_dict_: dictionary The dictionary containing the {ordinal number: category} pairs used to replace categories for every variable. """ def __init__(self, encoding_method = 'ordered', variables = None): if encoding_method not in ['ordered', 'arbitrary']: raise ValueError("encoding_method takes only values 'ordered' and 'arbitrary'") self.encoding_method = encoding_method self.variables = _define_variables(variables) def fit(self, X, y=None): """ Learns the numbers that should be used to replace the labels in each variable. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The training input samples. Can be the entire dataframe, not just seleted variables. y : Target. Can be None if selecting encoding_method = 'arbitrary'. Otherwise, needs to be passed when fitting the transformer. """ # brings the variables from the BaseEncoder super().fit(X, y) if self.encoding_method == 'ordered': if y is None: raise ValueError('Please provide a target (y) for this encoding method') temp = pd.concat([X, y], axis=1) temp.columns = list(X.columns)+['target'] self.encoder_dict_ = {} for var in self.variables: if self.encoding_method == 'ordered': t = temp.groupby([var])['target'].mean().sort_values(ascending=True).index elif self.encoding_method == 'arbitrary': t = X[var].unique() self.encoder_dict_[var] = {k:i for i, k in enumerate(t, 0)} if len(self.encoder_dict_)==0: raise ValueError('Encoder could not be fitted. Check that correct parameters and dataframe were passed during training') self.input_shape_ = X.shape return self class MeanCategoricalEncoder(BaseCategoricalTransformer): """ The MeanCategoricalEncoder() replaces categories by the mean of the target. For example in the variable colour, if the mean of the target for blue, red and grey is 0.5, 0.8 and 0.1 respectively, blue is replaced by 0.5, red by 0.8 and grey by 0.1. The Encoder will encode only categorical variables (type 'object'). A list of variables can be passed as an argument. If no variables are passed as argument, the encoder will only encode categorical variables (object type) and ignore the rest. The encoder first maps the categories to the numbers for each variable (fit). The encoder then transforms the categories to the mapped numbers (transform). Parameters ---------- variables : list, default=None The list of categorical variables that will be encoded. If None, the encoder will find and select all object type variables. Attributes ---------- encoder_dict_: dictionary The dictionary containing the {target mean: category} pairs used to replace categories for every variable """ def __init__(self, variables = None): self.variables = _define_variables(variables) def fit(self, X, y): """ Learns the numbers that should be used to replace the labels in each variable. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The training input samples. Can be the entire dataframe, not just seleted variables. y : Target """ # brings the variables from the BaseEncoder super().fit(X, y) if y is None: raise ValueError('Please provide a target (y) for this encoding method') temp = pd.concat([X, y], axis=1) temp.columns = list(X.columns)+['target'] self.encoder_dict_ = {} for var in self.variables: self.encoder_dict_[var] = temp.groupby(var)['target'].mean().to_dict() if len(self.encoder_dict_)==0: raise ValueError('Encoder could not be fitted. Check that correct parameters and dataframe were passed during training') self.input_shape_ = X.shape return self def transform(self, X): """ Replaces categories with the estimated numbers. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features]. The input samples. Returns ------- X_transformed : pandas dataframe of shape = [n_samples, n_features]. The dataframe containing categories replaced by numbers. """ # Check that the method fit has been called check_is_fitted(self, ['encoder_dict_']) # Check that the input is of the same shape as the training set passed # during fit. if X.shape[1] != self.input_shape_[1]: raise ValueError('Number of columns in dataset is different from train set used to fit the encoder') # encode labels X = X.copy() for feature in self.variables: X[feature+'_mean'] = X[feature].map(self.encoder_dict_[feature], na_action='ignore') if X[feature+'_mean'].isnull().sum() > 0: X[feature+'_mean'] = X[feature+'_mean'].fillna(0) warnings.warn("NaN values were introduced by the encoder due to labels in variable {} not present in the training set. Try using the RareLabelCategoricalEncoder.".format(feature) ) return X class MedianCategoricalEncoder(BaseCategoricalTransformer): """ The MedianCategoricalEncoder() replaces categories by the mean of the target. For example in the variable colour, if the medan of the target for blue, red and grey is 0.5, 0.8 and 0.1 respectively, blue is replaced by 0.5, red by 0.8 and grey by 0.1. The Encoder will encode only categorical variables (type 'object'). A list of variables can be passed as an argument. If no variables are passed as argument, the encoder will only encode categorical variables (object type) and ignore the rest. The encoder first maps the categories to the numbers for each variable (fit). The encoder then transforms the categories to the mapped numbers (transform). Parameters ---------- variables : list, default=None The list of categorical variables that will be encoded. If None, the encoder will find and select all object type variables. Attributes ---------- encoder_dict_: dictionary The dictionary containing the {target mean: category} pairs used to replace categories for every variable """ def __init__(self, variables = None): self.variables = _define_variables(variables) def fit(self, X, y): """ Learns the numbers that should be used to replace the labels in each variable. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The training input samples. Can be the entire dataframe, not just seleted variables. y : Target """ # brings the variables from the BaseEncoder super().fit(X, y) if y is None: raise ValueError('Please provide a target (y) for this encoding method') temp = pd.concat([X, y], axis=1) temp.columns = list(X.columns)+['target'] self.encoder_dict_ = {} for var in self.variables: self.encoder_dict_[var] = temp.groupby(var)['target'].median().to_dict() if len(self.encoder_dict_)==0: raise ValueError('Encoder could not be fitted. Check that correct parameters and dataframe were passed during training') self.input_shape_ = X.shape return self def transform(self, X): """ Replaces categories with the estimated numbers. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features]. The input samples. Returns ------- X_transformed : pandas dataframe of shape = [n_samples, n_features]. The dataframe containing categories replaced by numbers. """ # Check that the method fit has been called check_is_fitted(self, ['encoder_dict_']) # Check that the input is of the same shape as the training set passed # during fit. if X.shape[1] != self.input_shape_[1]: raise ValueError('Number of columns in dataset is different from train set used to fit the encoder') # encode labels X = X.copy() for feature in self.variables: X[feature+'_median'] = X[feature].map(self.encoder_dict_[feature], na_action='ignore') if X[feature+'_median'].isnull().sum() > 0: X[feature+'_median'] = X[feature+'_median'].fillna(0) warnings.warn("NaN values were introduced by the encoder due to labels in variable {} not present in the training set. Try using the RareLabelCategoricalEncoder.".format(feature) ) return X class RareLabelCategoricalEncoder(BaseCategoricalEncoder): """ The RareLabelCategoricalEncoder() groups rare / infrequent categories in a new category called "Rare". For example in the variable colour, if the percentage of observations for the categories magenta, cyan and burgundy are < 5 %, all those categories will be replaced by the new label "Rare". The Encoder will encode only categorical variables (type 'object'). A list of variables can be passed as an argument. If no variables are passed as argument, the encoder will only encode categorical variables (object type) and ignore the rest. The encoder first finds the frequent labels for each variable (fit). The encoder then groups the infrequent labels under the new label 'Rare' (transform). Parameters ---------- tol: float, default=0.05 the minimum frequency a label should have to be considered frequent and not be removed. n_categories: int, default=10 the minimum number of categories a variable should have in order for the encoder to find frequent labels. If the variable contains less categories, all of them will be considered frequent. variables : list, default=None The list of categorical variables that will be encoded. If None, the encoder will find and select all object type variables. Attributes ---------- encoder_dict_: dictionary The dictionary containg the frequent categories (that will be kept) for each variable. Categories not present in this list will be replaced by 'Rare'. """ def __init__(self, tol = 0.05, n_categories = 10, variables = None): if tol <0 or tol >1 : raise ValueError("tol takes values between 0 and 1") if n_categories < 0 or not isinstance(n_categories, int): raise ValueError("n_categories takes only positive integer numbers") self.tol = tol self.n_categories = n_categories self.variables = _define_variables(variables) def fit(self, X, y = None): """ Learns the frequent categories for each variable. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The training input samples. Can be the entire dataframe, not just seleted variables y : None There is no need of a target in a transformer, yet the pipeline API requires this parameter. You can leave y as None, or pass it as an argument. """ # brings the variables from the BaseEncoder super().fit(X, y) self.encoder_dict_ = {} for var in self.variables: if len(X[var].unique()) > self.n_categories: # if the variable has more than the indicated number of categories # the encoder will learn the most frequent categories t = pd.Series(X[var].value_counts() / np.float(len(X))) # non-rare labels: self.encoder_dict_[var] = t[t>=self.tol].index else: # if the total number of categories is smaller than the indicated # the encoder will consider all categories as frequent. self.encoder_dict_[var]= X[var].unique() self.input_shape_ = X.shape return self def transform(self, X): """ Groups rare labels under separate group 'Rare'. Parameters ---------- X : pandas dataframe of shape = [n_samples, n_features] The input samples. Returns ------- X_transformed : pandas dataframe of shape = [n_samples, n_features] The dataframe where rare categories have been grouped. """ # Check is fit had been called check_is_fitted(self, ['encoder_dict_']) # Check that the input is of the same shape as the one passed # during fit. if X.shape[1] != self.input_shape_[1]: raise ValueError('Number of columns in dataset is different from training set used to fit the encoder') X = X.copy() for feature in self.variables: X[feature] = np.where(X[feature].isin(self.encoder_dict_[feature]), X[feature], 'Rare') return X
[ "giuice@gmail.com" ]
giuice@gmail.com
e6ce7ed7be75cb36861a415589dd9916147d2c3b
17ed1b63dcc519c1ddf91d500e2b1e80c8cb338f
/Fixed_Income_Project/pricing_2.py
82ee24b37e9daa928d8f3fa6f694ec9230ef7465
[]
no_license
khorwei01/reinforcement
6834187814901b216317c2e2e29eed873a43b242
cfd2a151d1ecfae9f00ed658b45304ef0ff43046
refs/heads/master
2021-11-01T00:15:05.047005
2021-10-18T07:43:16
2021-10-18T07:43:16
238,147,407
0
0
null
null
null
null
UTF-8
Python
false
false
9,703
py
# -*- coding: utf-8 -*- """ FI part 4 @author: Archer """ import numpy as np import pandas as pd from enum import Enum from scipy.stats import norm from scipy.integrate import quad from scipy.misc import derivative from math import log from scipy.optimize import fsolve # discounting methods from part1 import OISdiscountFactorOutput, Fswapr from part2 import getSABR def sigmanN(n,N,K,m): N = N-n S = ParSwapRate(n,N,m) T = N alpha, rho, nu = getSABR(n, N-n) sigma_sabr = SABR(S, K, T, alpha, 0.9, rho, nu) return sigma_sabr def SABR(F, K, T, alpha, beta, rho, nu): X = K # if K is at-the-money-forward if abs(F - K) < 1e-12: numer1 = (((1 - beta)**2)/24)*alpha*alpha/(F**(2 - 2*beta)) numer2 = 0.25*rho*beta*nu*alpha/(F**(1 - beta)) numer3 = ((2 - 3*rho*rho)/24)*nu*nu VolAtm = alpha*(1 + (numer1 + numer2 + numer3)*T)/(F**(1-beta)) sabrsigma = VolAtm else: z = (nu/alpha)*((F*X)**(0.5*(1-beta)))*log(F/X) zhi = log((((1 - 2*rho*z + z*z)**0.5) + z - rho)/(1 - rho)) numer1 = (((1 - beta)**2)/24)*((alpha*alpha)/((F*X)**(1 - beta))) numer2 = 0.25*rho*beta*nu*alpha/((F*X)**((1 - beta)/2)) numer3 = ((2 - 3*rho*rho)/24)*nu*nu numer = alpha*(1 + (numer1 + numer2 + numer3)*T)*z denom1 = ((1 - beta)**2/24)*(log(F/X))**2 denom2 = (((1 - beta)**4)/1920)*((log(F/X))**4) denom = ((F*X)**((1 - beta)/2))*(1 + denom1 + denom2)*zhi sabrsigma = numer/denom return sabrsigma # import from part1 for reference def OISref(n): return OISdiscountFactorOutput(n) #OISref(3) def ParSwapRate(n,N,m=2): '''underlying swap LIBOR/ collateralized ''' # flt = sum([LiborRate(i-1,i)*OISref(i) for i in range(n+1,N+1,m)]) # fix = sum([1*OISref(i) for i in range(n+1,N+1,m)]) # return flt/fix return Fswapr(n,N-n) #ParSwapRate(5,15) class PayoffType(Enum): Call = 0 Put = 1 def IRR(S,Tenor,m=2): '''sum of IRR discounting swap should pay from n+1 yr, so adjust m is payment frequency S is par swap rate n,N are yrs starting swap and stop swap Note: 1. swap first payment start from n+1 2. by default start from 1 to m*N ''' comps = [1/m/(1+S/m)**i for i in range(1,Tenor*m+1)] return sum(comps) def IRR_1d(S,Tenor,m=2): '''derivative once of IRR ''' comps = [-i*(1/m**2)/(1+S/m)**(i+1) for i in range(1,Tenor*m+1)] return sum(comps) def IRR_2d(S,Tenor,m=2): '''derivative twice of IRR ''' comps = [(i*(i+1))*(1/m**3)/(1+S/m)**(i+2) for i in range(1,Tenor*m+1)] return sum(comps) def Black76(S, K, r, sigma, T, PayoffType=PayoffType.Call): '''real Black76 should go with F=S*np.exp(r*T) ''' d1 = (np.log(S/K)+(r+sigma**2/2)*T) / (sigma*np.sqrt(T)) d2 = d1 - sigma*np.sqrt(T) func = { PayoffType.Call: lambda : S*norm.cdf(d1) - K*np.exp(-r*T)*norm.cdf(d2), PayoffType.Put: lambda : S*norm.cdf(d1) - K*np.exp(-r*T)*norm.cdf(d2) - S + K*np.exp(-r*T) } return func[PayoffType]() def SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Call, m=2): ''' Df is Discount Factor from Tn(swap start time) to T0 S0 refers to par swap rate observed at T0 with payment from Tn to TN swapTenor is swap tenor n is time before swap signanN is a function ''' irr = IRR(S0, swapTenor, m) sigmaK = sigmanN(n=n,N=n+swapTenor,K=K,m=m) b76p = Black76(S0, K, 0, sigmaK, n, PayoffType=PayoffType) # take r as 0 return Df*irr*b76p def testSwaptionPricer(): # swaption price test case: forward swaption 5*10 Df = 1 # discount factor K = 0.1 # strike n=5 swapTenor=10 S0 = ParSwapRate(n,n+swapTenor) # par swap rate at T0 swaptionp = SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Call) print('Swaption price test-Call:',swaptionp) swaptionp = SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Put) print('Swaption price test-Put:',swaptionp) def CMSPrice(payoff, g_d1, g_d2, swapTenor, n, m=2): ''' payoff(or g) is a function with parameter K(par swap rate) h_d1 is first difference with K h_d2 is twice difference with K Df from 0 to n(when swap starts) ''' if (n%1 != 0) or (swapTenor%1 != 0): print('Do not support float numbers for now.') return 0 S0 = ParSwapRate(n,n+swapTenor,m) Df = OISref(n) # specify used formulas, used analytical formulas as much as possible IRR_cms = lambda K: IRR(K,swapTenor) IRR_d1_cms = lambda K: IRR_1d(K,swapTenor) IRR_d2_cms = lambda K: IRR_2d(K,swapTenor) # h = payoff / IRR # h = lambda K: payoff(K)/IRR_cms(K) # not used h_d1 = lambda K: ( g_d1(K) / IRR_cms(K) - IRR_d1_cms(K) * payoff(K) / IRR_cms(K)**2 ) h_d2 = lambda K: ( g_d2(K) / IRR_cms(K) - IRR_d2_cms(K) * payoff(K) / (IRR_cms(K)**2) - 2 * IRR_d1_cms(K) * g_d1(K) / (IRR_cms(K)**2) + 2 * IRR_d1_cms(K)**2 * payoff(K)/(IRR_cms(K)**3) ) # Df in swaption must be from Tn to T0 for swaption # from derivation, we know that K to swap is strike, but for payoff function, is par swap rate. swaption_payer = lambda K : SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Call, m=m) swaption_receiver = lambda K : SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Put, m=m) # within quad quad1 = lambda K : h_d2(K)*swaption_receiver(K) quad2 = lambda K : h_d2(K)*swaption_payer(K) # sum parts # p3 and p4 are intergals, which can have divergent results p1 = Df * payoff(S0) p2 = h_d1(S0)*(swaption_payer(S0)-swaption_receiver(S0)) p3 = quad(quad1, 0, S0)[0] p4 = quad(quad2, S0, np.inf)[0] return p1 + p2 + p3 + p4 def CMSCapletPrice(payoff, g_d1, g_d2, swapTenor, n, capletstrike, m=2): if (n%1 != 0) or (swapTenor%1 != 0): print('Do not support float numbers for now.') return 0 S0 = ParSwapRate(n,n+swapTenor,m) Df = OISref(n) # specify used formulas, used analytical formulas as much as possible IRR_cms = lambda K: IRR(K,swapTenor ) IRR_d1_cms = lambda K: IRR_1d(K,swapTenor ) IRR_d2_cms = lambda K: IRR_2d(K,swapTenor ) # h = payoff / IRR # h = lambda K: payoff(K)/IRR_cms(K) # not used h_d1 = lambda K: ( g_d1(K) / IRR_cms(K) - IRR_d1_cms(K) * payoff(K) / IRR_cms(K)**2 ) h_d2 = lambda K: ( g_d2(K) / IRR_cms(K) - IRR_d2_cms(K) * payoff(K) / (IRR_cms(K)**2) - 2 * IRR_d1_cms(K) * g_d1(K) / (IRR_cms(K)**2) + 2 * IRR_d1_cms(K)**2 * payoff(K)/(IRR_cms(K)**3) ) # Df in swaption must be from Tn to T0 for swaption # from derivation, we know that K to swap is strike, but for payoff function, is par swap rate. swaption_payer = lambda K : SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Call, m=m) swaption_receiver = lambda K : SwaptionPrice(Df, S0, K, swapTenor, n, PayoffType=PayoffType.Put, m=m) # within quad quad1 = lambda K : h_d2(K)*swaption_receiver(K) quad2 = lambda K : h_d2(K)*swaption_payer(K) # sum parts # p3 and p4 are intergals, which can have divergent results if capletstrike < S0: p1 = Df * payoff(S0) p2 = h_d1(capletstrike)*swaption_receiver(capletstrike) p3 = quad(quad1, capletstrike, S0)[0] p4 = quad(quad2, S0, np.inf)[0] else: p1 = 0 p2 = h_d1(capletstrike)*swaption_payer(capletstrike) p3 = 0 p4 = quad(quad2, capletstrike, np.inf)[0] pcheck = quad(quad1, 0, capletstrike)[0] print('Caplet part2 and 3', [p2,pcheck]) print('h1d',h_d1(capletstrike),'\nh2d',h_d2(capletstrike)) print('rec',round(swaption_receiver(capletstrike),8)) return p1 + p2 + p3 + p4 class backup(): # derivative by derivative func IRR_d1_cms = lambda K: derivative(IRR_cms, K, dx=0.001 ,n=1) IRR_d2_cms = lambda K: derivative(IRR_cms, K, dx=0.001 ,n=2) # normal case for normal CMS paying K payoff = lambda K: K g_d1 = lambda K: 1 g_d2 = lambda K: 0 def peng(): m = 2 N = 15 n = 5 swapTenor = N-n # payoff equations payoff = lambda K: K g_d1 = lambda K: 1 g_d2 = lambda K: 0 # S0 = ParSwapRate(n,n+swapTenor) PV = CMSPrice(payoff, g_d1, g_d2, swapTenor, n, m) print('Q1: CMS PV:',PV) Df = OISref(n) print('CMS rate:',PV/Df) def part4(n=5, N=15): p=4 q=2 swapTenor = N-n # payoff equations payoff = lambda K: K**(1/p) - 0.04**(1/q) g_d1 = lambda K: (1/p)*K**(1/p-1) g_d2 = lambda K: 1/p*(1/p-1)*K**(1/p-2) m=2 # S0 = ParSwapRate(n,n+swapTenor) PV = CMSPrice(payoff, g_d1, g_d2, swapTenor, n, m) print('Q1: CMS PV:',round(PV,8)) Df = OISref(n) print('CMS rate:',round(PV/Df,8)) S0 = ParSwapRate(n,n+swapTenor) print('S0',S0) capletstrike = fsolve(payoff,0)[0] print('caplet strike', capletstrike) PVop = CMSCapletPrice(payoff, g_d1, g_d2, swapTenor, n, capletstrike, m) print('Q2: CMS Caplet PV:',round(PVop,8)) print('difference between Option - CMS', PVop-PV) if __name__ == '__main__': # so far cannot support n and N flt numbers part4(n=5, N=15)
[ "noreply@github.com" ]
noreply@github.com
cb1e1c4fd0adabebcd87bc33eefe453ec2df48fa
942ee5e8d54e8ebe9c5c841fbfdd1da652946944
/1001-1500/1443.Minimum Time to Collect All Apples in a Tree.py
e8ae7ff0deadce1de133f2d3d5feb31d43fde59a
[]
no_license
kaiwensun/leetcode
0129c174457f32887fbca078fb448adce46dd89d
6b607f4aae3a4603e61f2e2b7480fdfba1d9b947
refs/heads/master
2023-08-31T07:30:50.459062
2023-08-27T07:59:16
2023-08-27T07:59:16
57,526,914
69
9
null
2023-08-20T06:34:41
2016-05-01T05:37:29
Python
UTF-8
Python
false
false
728
py
from collections import defaultdict class Solution(object): def minTime(self, n, edges, hasApple): """ :type n: int :type edges: List[List[int]] :type hasApple: List[bool] :rtype: int """ graph = defaultdict(list) for edge in edges: graph[edge[0]].append(edge[1]) graph[edge[1]].append(edge[0]) visited = set() def dfs(root): res = 0 if root not in visited: visited.add(root) for nbr in graph[root]: res += dfs(nbr) if res or hasApple[root]: res += 2 return res return max(0, dfs(0) - 2)
[ "noreply@github.com" ]
noreply@github.com
d874e4f137ec4de13520364ebc883e120ec966a1
7f19038537cfafb2c725973d36ae0d1ae767995f
/Labs/lab10/lab10.py
378f26a0b28368e93b796d5f88913dbfcea81dc5
[]
no_license
aaeissa/UMBC-CS-201
f532684734543dd7a16698b61f1824443dfeb23e
c07a97469fc9c3a3b91e5228880abeb49918691f
refs/heads/master
2021-01-13T12:49:50.373730
2016-10-29T18:30:52
2016-10-29T18:30:52
72,298,768
0
0
null
null
null
null
UTF-8
Python
false
false
1,082
py
def convertToDict(info): stateFile = info.readlines() myDict = {} states = "" for line in stateFile: states = line.strip() key, value = states.split(",") myDict[key] = value return(myDict) def main(): print("Welcomet to the State Abbreviator.") info = open("states.txt", "r") myDict = convertToDict(info) userInput = "" while userInput != "exit" and userInput != "Exit": userInput = input("Please enter the state to abbreviate (list to get list and exit to exit): ").strip() if userInput == "list": for key in sorted(myDict): print(key) else: fullName = myDict.keys() abbreviation = myDict.values() if userInput in fullName: print("The abbreviation of the state: ", userInput, "is", myDict[userInput]) else: print("Sorry. That is not a state.") print("Thank you for using the State Abbreviator.") info.close() main()
[ "ahmedeissa33@gmail.com" ]
ahmedeissa33@gmail.com
7afc92885276c205066adf4255d7d46f5f5e60fb
30f9e6a5f649b52b5d1a0ae606f44764038f6028
/pai-fs/fsimpl/Errors.py
0ad8850b971ccf904dbdae10d1b94181ddf962f7
[ "MIT", "LicenseRef-scancode-generic-cla" ]
permissive
unixcrh/pai
97d6ffe6ff16eecd91a8e5222c7f4b33efddb1aa
5c1836edc1db6e7d3b23ad2416fec309d48fc677
refs/heads/master
2021-04-15T09:12:25.198703
2018-03-26T09:40:39
2018-03-26T09:40:39
126,845,731
1
0
MIT
2018-03-26T14:59:16
2018-03-26T14:59:15
null
UTF-8
Python
false
false
1,449
py
# Copyright (c) Microsoft Corporation # All rights reserved. # # MIT License # # Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated # documentation files (the "Software"), to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and # to permit persons to whom the Software is furnished to do so, subject to the following conditions: # The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED *AS IS*, WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING # BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, # DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. class FsException(Exception): def __init__(self, msg=str()): self.msg = msg super(FsException, self).__init__(self.msg) class BadConnection(FsException): pass class Unauthorized(FsException): pass class FileNotFound(FsException): pass class PathNotEmpty(FsException): pass
[ "v-yifxio@microsoft.com" ]
v-yifxio@microsoft.com
0866ea3ae0e864e61eef207ffc5ef2936d10b99e
9e6bab4664f9275896b510209a83d5917b4746e1
/apps/moca/migrations/0003_merchantprofile.py
7d57bdec7486b538d9b0c963255d7ab013ee3543
[]
no_license
manticarodrigo/moca
4a25215715a11b0a85d0ce65388b35df8c799eed
d8cce8dc5c88e8dc944bb7c5c89e11a86e009e03
refs/heads/master
2023-08-03T08:25:52.856943
2019-12-04T22:20:00
2019-12-04T22:20:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
717
py
# Generated by Django 2.1.10 on 2019-12-04 18:46 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('moca', '0002_device_auth_token'), ] operations = [ migrations.CreateModel( name='MerchantProfile', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('stripe_user_id', models.CharField(max_length=30)), ('user', models.OneToOneField(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), ]
[ "manticarodrigo@gmail.com" ]
manticarodrigo@gmail.com
8130572a3731398674da676762ba7c5a6251ef9f
79a23f55af0e8d4898714b7e210cadc88033a390
/website/views.py
eae6187e75e7acb17fd508e9063dcc70f47de1c6
[]
no_license
luoyueyaoqing/forum-website
be9bc7802d83a58abc3f802469287b3c2e11f44b
d5fbb14adf3a3e75f70df8f6eed397665f9c2d84
refs/heads/master
2020-03-25T10:01:16.533106
2018-08-15T08:48:55
2018-08-15T08:48:55
143,681,923
2
1
null
null
null
null
UTF-8
Python
false
false
3,756
py
from django.shortcuts import render, HttpResponseRedirect, redirect, HttpResponse from .models import User, Article, Plate, Comment from django.contrib import messages from django.contrib.auth import login, authenticate, logout from django.contrib.auth.forms import AuthenticationForm from django.contrib.auth.decorators import login_required def index_register(request): if request.method == "POST": username = request.POST.get('username') password1 = request.POST.get('password1') password2 = request.POST.get('password2') if not User.objects.filter(username=username).exists(): if password1 == password2: User.objects.create_user(username=username, password=password1) messages.success(request, '注册成功') return redirect(to='login') else: messages.warning(request, '两次密码输入不一致') else: messages.warning(request, "账号已存在") return render(request, 'register.html') def index_login(request): next_url = request.GET.get('next') if request.method == "POST": form = AuthenticationForm(data=request.POST) if form.is_valid(): login(request, form.get_user()) if next_url: return redirect(next_url) return redirect('index') return HttpResponseRedirect(request.get_full_path()) return render(request, 'login.html', {'next_url': next_url}) def index_logout(request): logout(request) return redirect(to=index) def index(request): plates = Plate.objects.all() return render(request, 'index.html', {'plates': plates}) def articles(request, id): plate = Plate.objects.get(id=id) articles = Article.objects.all() return render(request, 'articles.html', locals()) def detail(request, id): article = Article.objects.get(id=id) return render(request, 'detail.html', {'article': article}) @login_required def add_article(request, id): if request.method == "POST": plate = Plate.objects.get(id=id) title = request.POST.get('title') content = request.POST.get('content') article = Article.objects.create(title=title, content=content, author=request.user, column=plate) return redirect(to='articles', id=id) else: return render(request, 'add_article.html') @login_required def comment(request, id): article = Article.objects.get(id=id) content = request.POST.get('content') # user = request.user # Comment.objects.create(content=content, user=user, article=article) article.comment_this(user=request.user, content=content) messages.success(request, '评论成功') return redirect(to='detail', id=article.id) @login_required def edit(request, id): article = Article.objects.get(id=id) if request.method == "POST": article.title = request.POST.get('title') article.content = request.POST.get('content') article.save() return redirect(to=detail, id=id) return render(request, 'edit.html', {'article': article}) @login_required def del_article(request, id): if request.method == "GET": article = Article.objects.get(id=id) column = article.column article.delete() return redirect(to=articles, id=column.id) @login_required def del_comment(request, id): if request.method == "GET": comment = Comment.objects.get(id=id) comment.delete() article = comment.article messages.success(request, "评论已删除") return redirect(to=detail, id=article.id)
[ "shmilyfae@163.com" ]
shmilyfae@163.com
c2f2d9873572b84a36f2345329ebd77f92a88cbe
98e1716c1c3d071b2fedef0ac029eb410f55762c
/part15-statistical-thinking-1/No04-Bee-swarm-plot.py
0b503f7631dcaaedd5a7afe2edbda8d651de8a7c
[]
no_license
iamashu/Data-Camp-exercise-PythonTrack
564531bcf1dff119949cbb75e1fd63d89cb2779f
c72a4e806494f0e263ced9594597dc8882c2131c
refs/heads/master
2020-07-22T00:23:12.024386
2019-04-12T09:24:42
2019-04-12T09:24:42
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,212
py
#Bee swarm plot ''' Make a bee swarm plot of the iris petal lengths. Your x-axis should contain each of the three species, and the y-axis the petal lengths. A data frame containing the data is in your namespace as df. For your reference, the code Justin used to create the bee swarm plot in the video is provided below: _ = sns.swarmplot(x='state', y='dem_share', data=df_swing) _ = plt.xlabel('state') _ = plt.ylabel('percent of vote for Obama') plt.show() In the IPython Shell, you can use sns.swarmplot? or help(sns.swarmplot) for more details on how to make bee swarm plots using seaborn. Instructions In the IPython Shell, inspect the DataFrame df using df.head(). This will let you identify which column names you need to pass as the x and y keyword arguments in your call to sns.swarmplot(). Use sns.swarmplot() to make a bee swarm plot from the DataFrame containing the Fisher iris data set, df. The x-axis should contain each of the three species, and the y-axis should contain the petal lengths. Label the axes. Show your plot. ''' # code sns.swarmplot(x='species', y='petal length (cm)', data=df) # Label the axes plt.xlabel('species') plt.ylabel('petal length (cm)') # Show the plot plt.show()
[ "beiran@hotmail.com" ]
beiran@hotmail.com
f442cd7e453fbda6168780d39defdc6005872da8
576e36669d9012c244fe2f777314ba1a6d903277
/ChefAndPizza_problem_PCJ18G.py
fe82a5fe7766e9a759c1bf080cb7195945214b8b
[]
no_license
OlhaKuzaka/algorithms_python
cf26658e1f70491eda4cb636269c0373aed9f76d
14edacc8f0c0040fc95e9cae0e88515ac5293985
refs/heads/master
2020-03-26T10:37:50.018355
2018-08-15T05:03:23
2018-08-15T05:03:23
144,808,431
0
0
null
null
null
null
UTF-8
Python
false
false
1,376
py
# All submissions for this problem are available. # Chef's pizza is the tastiest pizza to exist, and the reason for that is his special, juicy homegrown tomatoes. # # Tomatoes can be grown in rectangular patches of any side lengths. However, Chef only has a limited amount of land. # # Consider the entire town of Chefville to be consisting of cells in a rectangular grid of positive coordinates. # # Chef own all cells (x,y) # that satisfy x∗y≤N # As an example if N=4 # , Chef owns the following cells: # # (1,1),(1,2),(1,3),(1,4),(2,1),(2,2),(3,1),(4,1) # Chef can only grow tomatoes in rectangular patches consisting only of cells which belong to him. # Also, if he uses a cell, he must use it entirely. He cannot use only a portion of it. # # Help Chef find the number of unique patches of rectangular land that he can grow tomatoes in! # Since this number can be very large, output it modulo 1000000007 # . # # Input: # The first line of the input contains T # , the number of test cases. # The next T # lines of input contains one integer N # . def num_rectangles(x, y): return x*(x+1) * y*(y+1)/4 if __name__ == '__main__': n = 10000000000 total = 0 for i in range(1, n+1): edge = int(n/i) total += num_rectangles(i, edge) if i-1 != 0: total -= num_rectangles(i-1, edge) print(total % 1000000007)
[ "o.kuzka@gmail.com" ]
o.kuzka@gmail.com
8c02b65c424703478a511dc63b5449224c33868c
d2d9ff977d8eb21420d6f3dc5f9476850ebcc5e9
/ABC_SMC_exponential_F5.py
1d9a14cf7871f9587428b409151c8bc433fd8962
[ "MIT" ]
permissive
GiuliaBell/ABC_SMC_codes
45d85fd22c2b36058f6a3b4cf66b0bc590209781
6587190456c9465a1cb056916d7502a1ea8d6330
refs/heads/main
2023-04-17T20:01:20.503306
2022-03-08T13:09:42
2022-03-08T13:09:42
467,183,547
0
0
null
null
null
null
UTF-8
Python
false
false
12,223
py
# Pyhton code to run the ABC-SMC algorithm to parametrise the exponential model with cell generations # making use of F5 T cells data. # Reference: "Approximate Bayesian Computation scheme for parameter inference and model selection # in dynamical systems" by Toni T. et al. (2008). # Import the required modules. import numpy as np from scipy.linalg import expm G = 11 # Total number of generations. dist_gen = 6 # Used to define generation 5+. n_pars = 4 # Number of parameters in the exponential model. N0 = 1 # Number of stages in generation 0. N1 = 1 # Number of stages all generations but 0. # Reading the data. data = np.loadtxt("data_F5.txt") std_dev = np.loadtxt("std_dev_F5.txt") # Define the time points (unit of hours). t2 = np.array([72, 96, 120, 144, 168, 240, 288, 432]) # Define the exponential model with cell generations. def diag(g,N,l,m): if g < 5: return(np.diag([-(l[g]+m[g])]*N) + np.diag([l[g]]*(N-1),-1)) else: return(np.diag([-(l[5]+m[5])]*N) + np.diag([l[5]]*(N-1),-1)) def matrix(N0,N1,l,m): M = np.zeros((N0+(G-1)*N1, N0+(G-1)*N1)) M[0:N0,0:N0] = diag(0,N0,l,m) for i in range(1,G): M[N0+(i-1)*N1:N0+i*N1,N0+(i-1)*N1:N0+i*N1] = diag(i,N1,l,m) M[N0,N0-1] = 2*l[0] for i in range(1,G-1): if i < 5: M[N0+i*N1,N0+i*N1-1] = 2*l[i] else: M[N0+i*N1,N0+i*N1-1] = 2*l[5] return(M) def exp_matrix(N0,N1,inits,times,l,m): output = np.zeros((len(times),N0+N1*(G-1))) A = matrix(N0,N1,l,m) for i in range(len(times)): sol = np.dot(expm(A*times[i]),inits) output[i] = sol return output.T # Define the functions to use in the ABC-SMC algorithm to generate the first epsilon, to run the first iteration # and to run all the other iterations. # As it may be difficult to decide on a reasonably large value of epsilon to use at the first iteration, # we defined the function below to generate it. def generate_eps1(nn,rr): # Empty array to store the distance. results = np.empty((0)) # Empty array to store the accepted parameters. params = np.empty((0,n_pars)) for run in range(nn*rr): # Sample the parameters from uniform prior distributions. l0, lambd = 10**np.random.uniform(-3,1,2) l = np.array([lambd for _ in range(dist_gen)]) l[0] = l0 alpha = 10**np.random.uniform(-5,-1) m = np.zeros(dist_gen) for i in range(dist_gen): m[i] = alpha*i C0 = 10**np.random.uniform(4,6) inits = np.zeros((N0+(G-1)*N1)) inits[0] = C0 # Run the model to compute the expected number of cells in each generation. generations = [[] for _ in range(dist_gen)] modelexp = exp_matrix(N0,N1,inits,t2,l,m) s0 = sum(modelexp[0:N0]) generations[0].append(s0) for i in range(1,dist_gen): if i < 5: s = sum(modelexp[N0+(i-1)*N1:N0+i*N1]) generations[i].append(s) else: s = sum(modelexp[N0+(i-1)*N1:N0+(G-1)*N1]) generations[i].append(s) # Compute the distance between the model predictions and the experimental data. generationsravel = np.ravel(generations) dataravel = np.ravel(data) std_ravel = np.ravel(std_dev) distance = np.sqrt(np.sum(((generationsravel-dataravel)/std_ravel)**2)) results = np.hstack((results, distance)) params = np.vstack((params, np.hstack((C0,l0,lambd,alpha)))) # Compute epsilon to use at the first iteration. epsilon = np.median(results) return epsilon # Define the function for the first iteration of ABC-SMC in which the parameters are sampled # from the uniform prior distributions. def iteration1(nn): # Empty array to store the distance. results = np.empty((0,1)) # Empty array to store the accepted parameters. params = np.empty((0,n_pars)) number = 0 # Counter for the sample size. truns = 0 # Counter for the total number of runs. while number < nn: truns+=1 # Sample the parameters from uniform prior distributions. l0, lambd = 10**np.random.uniform(-3,1,2) l = np.array([lambd for _ in range(dist_gen)]) l[0] = l0 alpha = 10**np.random.uniform(-5,-1) m = np.zeros(dist_gen) for i in range(dist_gen): m[i] = alpha*i C0 = 10**np.random.uniform(4,6) inits = np.zeros((N0+(G-1)*N1)) inits[0] = C0 pars=np.hstack((C0,l0,lambd,alpha)) # Run the model to compute the expected number of cells in each generation. generations = [[] for _ in range(dist_gen)] modelexp = exp_matrix(N0,N1,inits,t2,l,m) s0 = sum(modelexp[0:N0]) generations[0].append(s0) for i in range(1,dist_gen): if i < 5: s = sum(modelexp[N0+(i-1)*N1:N0+i*N1]) generations[i].append(s) else: s = sum(modelexp[N0+(i-1)*N1:N0+(G-1)*N1]) generations[i].append(s) # Compute the distance between the model predictions and the experimental data. generationsravel = np.ravel(generations) dataravel = np.ravel(data) std_ravel = np.ravel(std_dev) distance = np.sqrt(np.sum(((generationsravel-dataravel)/std_ravel)**2)) # If the distance is less than epsilon, store the parameters values and increase by one the counter for # the sample size. if distance < eps1: number+=1 results = np.vstack((results, distance)) params = np.vstack((params, pars)) # Compute the weight for each accepted parameter set - at iteration 1, parameter sets have equal weight. weights = np.empty((0,1)) for i in range(nn): weights = np.vstack((weights,1/nn)) # Return the results: distance, accepted parameters, weights and total number of runs. return [np.hstack((results,params,weights)), truns] # Function for the other iterations of the ABC-SMC algorithm, where the parameter values are sampled # from the posterior distributions of the previous iteration. def other_iterations(nn,it): # Compute uniform areas to sample within in order to perturb the parameters. ranges = [] for i in range(n_pars): r1 = np.max(np.log10(ABC_runs[it][:,i+1])) - np.min(np.log10(ABC_runs[it][:,i+1])) ranges.append(r1) ranges_arr = np.asarray(ranges) sigma = 0.1*ranges_arr # Define epsilon as median of the accepted distance values from previous iteration. epsilon = np.median(ABC_runs[it][:,0]) # To use when sampling the new parameters. p_list = [i for i in range(nn)] # Define upper and lower bounds of the prior distributions for each parameter in the model. lower_bounds = np.hstack((10**4,10**(-3),10**(-3),10**(-5))) upper_bounds = np.hstack((10**6,10,10,10**(-1))) # Empty array to store the distance. results = np.empty((0)) # Empty array to store accepted parameters. params = np.empty((0,n_pars)) # Empty array to store the prior samples. priors_abc = np.empty((0,n_pars)) # Empty array to store the weights. weights_arr = np.empty((0)) number = 0 # Counter for the sample size. truns = 0 # Counter for the total number of runs. while number < nn: truns+=1 check = 0 # The following while loop is to sample the parameters from the posterior distributions of the previous # iteration. Then the parameters are perturbed making use of a uniform perturbation kernel. # If the new parameters lie within the initial prior ranges, they are used to obtaining model predictions, # otherwise they are sampled again. while check < 1: # Randomly choose a parameter set from the posterior obtained from the previous iteration. choice = np.random.choice(p_list,1,p=ABC_runs[it][:,n_pars+1]) prior_sample = ABC_runs[it][:,range(1,n_pars+1)][choice] # Generate new parameters through perturbation. parameters = [] for i in range(n_pars): lower = np.log10(prior_sample[0,i])-sigma[i] upper = np.log10(prior_sample[0,i])+sigma[i] pars = np.random.uniform(lower,upper) parameters.append(10**pars) # Check that the new parameters lie within the initial prior ranges. check_out = 0 for ik in range(n_pars): if parameters[ik] < lower_bounds[ik] or parameters[ik] > upper_bounds[ik]: check_out = 1 if check_out == 0: check+=1 C0 = float(parameters[0]) l0, lambd = parameters[1:3] l = np.array([lambd for _ in range(dist_gen)]) l[0] = l0 m = np.zeros(dist_gen) for i in range(dist_gen): m[i] = i*parameters[3] inits = np.zeros((N0+(G-1)*N1)) inits[0] = C0 # Run the model to compute the expected number of cells in each generation. generations = [[] for _ in range(dist_gen)] modelexp = exp_matrix(N0,N1,inits,t2,l,m) s0 = sum(modelexp[0:N0]) #it stops at N0-1 generations[0].append(s0) for i in range(1,dist_gen): if i < 5: s = sum(modelexp[N0+(i-1)*N1:N0+i*N1]) generations[i].append(s) else: s = sum(modelexp[N0+(i-1)*N1:N0+(G-1)*N1]) generations[i].append(s) # Compute the distance between the model predictions and the experimental data. generationsravel = np.ravel(generations) dataravel = np.ravel(data) std_ravel = np.ravel(std_dev) distance = np.sqrt(np.sum(((generationsravel-dataravel)/std_ravel)**2)) # If the distance is less than epsilon, store the parameters values and increase by one the counter for # the sample size. if distance < epsilon: number+=1 # Compute the weights for the accepted parameter set. denom_arr = [] for j in range(nn): weight = ABC_runs[it][j,n_pars+1] params_row = ABC_runs[it][j,1:n_pars+1] boxs_up = [] boxs_low = [] for i in range(n_pars): boxs_up.append(np.log10(params_row[i]) + sigma[i]) boxs_low.append(np.log10(params_row[i]) - sigma[i]) outside = 0 for i in range(n_pars): if np.log10(parameters[i]) < boxs_low[i] or np.log10(parameters[i]) > boxs_up[i]: outside = 1 if outside == 1: denom_arr.append(0) else: denom_arr.append(weight*np.prod(1/(2*sigma))) weight_param = 1/np.sum(denom_arr) weights_arr = np.hstack((weights_arr,weight_param)) results = np.hstack((results, distance)) params = np.vstack((params, parameters)) priors_abc = np.vstack((priors_abc, prior_sample)) # Normalise the weights. weights_arr2 = weights_arr/np.sum(weights_arr) weights_arr3 = np.reshape(weights_arr2, (nn,1)) # Return the results: distance, accepted parameters, weights and total number of runs. return [np.hstack((np.reshape(results,(nn,1)),params,weights_arr3)),epsilon,truns] # Sample size for the ABC-SMC. sample_size = 10000 # Number of iterations to run. num_iters = 7 # To generate the first value of epsilon. eps1 = generate_eps1(sample_size,1) Epsilons = [eps1] # Run the first iteration of ABC-SMC. first_output = iteration1(sample_size) ABC_runs = [first_output[0]] # Run all the other iterations of ABC-SMC. for iterat in range(num_iters): run = other_iterations(sample_size,iterat) ABC_runs.append(run[0]) Epsilons.append(run[1]) # Save the results as a text file. np.savetxt('Posterior_F5_exponential.txt', ABC_runs[num_iters])
[ "95767138+GiuliaBell@users.noreply.github.com" ]
95767138+GiuliaBell@users.noreply.github.com
15a3b3cd280903839a7706abb9e05c8e6621ad2a
837877c72a803fc62eea396e466dff36cffa5028
/dnnlib/submission/internal/__init__.py
52a7fce687d3906d061a6a76834e222143555b9b
[]
no_license
tantantetetao/StyleGANv2-Re-Implementation
55d6c9e98e2d98450734937133348acb197bf5d8
bfa0c410073863ab199cc5e5fbafb3dd2da35ea9
refs/heads/master
2022-12-12T23:32:53.423432
2020-09-14T07:26:27
2020-09-14T07:26:27
null
0
0
null
null
null
null
UTF-8
Python
false
false
281
py
# Copyright (c) 2019, NVIDIA Corporation. All rights reserved. # # This work is made available under the Nvidia Source Code License-NC. # To view a copy of this license, visit # https://nvlabs.github.io/stylegan2/license.html # Cheng-Bin Jin re-implementation. from . import local
[ "jinchengbin@huya.com" ]
jinchengbin@huya.com
59a23d259eca385a7a14f82cc06da91203ce8f7c
9e30d4e6e14c5a57d99f80b28b59ee42f23fe4eb
/Listas.py
c1caff2066d60c415e7c5c3aedf552bd7ca108f7
[]
no_license
lukeencinas/EjerciciosExtraPython
408b1208f526a11a4ae214fd8c15b3fd31216626
e75378d8b0fb12684c5cfeba396a50d3ec030c1d
refs/heads/master
2020-05-09T18:25:33.295651
2019-04-14T17:04:31
2019-04-14T17:04:31
181,341,719
0
0
null
null
null
null
UTF-8
Python
false
false
623
py
#1) e1 = ["F. de Computadores", "Visualizacion","Fisica","Quimica", "Historia", "Lengua"] # print(e1) #2) e2 = ["F. de Computadores", "Visualizacion","Fisica","Quimica", "Historia", "Lengua"] # for i in e2: # print("Yo estudio: "+i) #3) e3 = ["F. de Computadores", "Visualizacion","Fisica"] # nota = [] # for i in e3: # nota.append(input('¿Que has sacado en '+ i + '? ')) # for i in range(len(e3)): # print ("En la asignatura ",e3[i]," Has sacado ",nota[i]) e4 = [] for i in range(6): e4.append(int(input("¿Cuales son los numeros de la loteria? "))) e4.sort() print("Los numeros en orden son: "+ str(e4))
[ "noreply@github.com" ]
noreply@github.com
c03967857b3abb3a4db4df537c2c4342ac393b68
d41d18d3ea6edd2ec478b500386375a8693f1392
/plotly/validators/scatter/marker/line/_width.py
108770c589b4e4605b6ff605e20647ef337325b7
[ "MIT" ]
permissive
miladrux/plotly.py
38921dd6618650d03be9891d6078e771ffccc99a
dbb79e43e2cc6c5762251537d24bad1dab930fff
refs/heads/master
2020-03-27T01:46:57.497871
2018-08-20T22:37:38
2018-08-20T22:37:38
145,742,203
1
0
MIT
2018-08-22T17:37:07
2018-08-22T17:37:07
null
UTF-8
Python
false
false
465
py
import _plotly_utils.basevalidators class WidthValidator(_plotly_utils.basevalidators.NumberValidator): def __init__( self, plotly_name='width', parent_name='scatter.marker.line', **kwargs ): super(WidthValidator, self).__init__( plotly_name=plotly_name, parent_name=parent_name, array_ok=True, edit_type='style', min=0, role='style', **kwargs )
[ "adam.kulidjian@gmail.com" ]
adam.kulidjian@gmail.com
aaab34fc3d46736a9c6abdb2195fc43a0ba0f6f4
75f48cb18eb83fd4f9fd1c7dd0543fcaaf781342
/members/migrations/0029_profile.py
e7b48cc47f392bbf73ff5bb36412d5885b6a87dc
[]
no_license
trouvaay/trouvaay
f28a879e60de6212cbf30904c4471158c3907ebb
5440928238428c9e60be61e75910ad8d78866426
refs/heads/master
2021-03-27T10:11:09.602842
2015-06-05T22:06:28
2015-06-05T22:06:28
27,945,386
0
0
null
null
null
null
UTF-8
Python
false
false
817
py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations import localflavor.us.models from django.conf import settings class Migration(migrations.Migration): dependencies = [ ('members', '0028_auto_20150220_1922'), ] operations = [ migrations.CreateModel( name='Profile', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('phone', localflavor.us.models.PhoneNumberField(default=b'', max_length=20, blank=True)), ('authuser', models.OneToOneField(related_name='profile', to=settings.AUTH_USER_MODEL)), ], options={ }, bases=(models.Model,), ), ]
[ "sergey@lanasoft.net" ]
sergey@lanasoft.net
1161c6ec01e8bf8124199a123fc850feb16f7924
27c94d7e040902d3cdadd5862b15e67ec2ee4b6e
/exps/NAS-Bench-201-algos/DARTS-V1.py
67441af82a7bc2f760fa028163eb4ca9c8887773
[ "MIT" ]
permissive
D-X-Y/AutoDL-Projects
8a0779a7710d809af2b052787928d8d34c14d0d9
f46486e21b71ae6459a700be720d7648b5429569
refs/heads/main
2023-08-13T10:53:49.550889
2022-04-24T22:18:16
2022-04-24T22:18:16
168,538,768
989
197
MIT
2022-04-24T22:16:21
2019-01-31T14:30:50
Python
UTF-8
Python
false
false
15,785
py
################################################## # Copyright (c) Xuanyi Dong [GitHub D-X-Y], 2020 # ######################################################## # DARTS: Differentiable Architecture Search, ICLR 2019 # ######################################################## import sys, time, random, argparse from copy import deepcopy import torch from pathlib import Path from xautodl.config_utils import load_config, dict2config, configure2str from xautodl.datasets import get_datasets, get_nas_search_loaders from xautodl.procedures import ( prepare_seed, prepare_logger, save_checkpoint, copy_checkpoint, get_optim_scheduler, ) from xautodl.utils import get_model_infos, obtain_accuracy from xautodl.log_utils import AverageMeter, time_string, convert_secs2time from xautodl.models import get_cell_based_tiny_net, get_search_spaces from nas_201_api import NASBench201API as API def search_func( xloader, network, criterion, scheduler, w_optimizer, a_optimizer, epoch_str, print_freq, logger, gradient_clip, ): data_time, batch_time = AverageMeter(), AverageMeter() base_losses, base_top1, base_top5 = AverageMeter(), AverageMeter(), AverageMeter() arch_losses, arch_top1, arch_top5 = AverageMeter(), AverageMeter(), AverageMeter() network.train() end = time.time() for step, (base_inputs, base_targets, arch_inputs, arch_targets) in enumerate( xloader ): scheduler.update(None, 1.0 * step / len(xloader)) base_targets = base_targets.cuda(non_blocking=True) arch_targets = arch_targets.cuda(non_blocking=True) # measure data loading time data_time.update(time.time() - end) # update the weights w_optimizer.zero_grad() _, logits = network(base_inputs) base_loss = criterion(logits, base_targets) base_loss.backward() if gradient_clip > 0: torch.nn.utils.clip_grad_norm_(network.parameters(), gradient_clip) w_optimizer.step() # record base_prec1, base_prec5 = obtain_accuracy( logits.data, base_targets.data, topk=(1, 5) ) base_losses.update(base_loss.item(), base_inputs.size(0)) base_top1.update(base_prec1.item(), base_inputs.size(0)) base_top5.update(base_prec5.item(), base_inputs.size(0)) # update the architecture-weight a_optimizer.zero_grad() _, logits = network(arch_inputs) arch_loss = criterion(logits, arch_targets) arch_loss.backward() a_optimizer.step() # record arch_prec1, arch_prec5 = obtain_accuracy( logits.data, arch_targets.data, topk=(1, 5) ) arch_losses.update(arch_loss.item(), arch_inputs.size(0)) arch_top1.update(arch_prec1.item(), arch_inputs.size(0)) arch_top5.update(arch_prec5.item(), arch_inputs.size(0)) # measure elapsed time batch_time.update(time.time() - end) end = time.time() if step % print_freq == 0 or step + 1 == len(xloader): Sstr = ( "*SEARCH* " + time_string() + " [{:}][{:03d}/{:03d}]".format(epoch_str, step, len(xloader)) ) Tstr = "Time {batch_time.val:.2f} ({batch_time.avg:.2f}) Data {data_time.val:.2f} ({data_time.avg:.2f})".format( batch_time=batch_time, data_time=data_time ) Wstr = "Base [Loss {loss.val:.3f} ({loss.avg:.3f}) Prec@1 {top1.val:.2f} ({top1.avg:.2f}) Prec@5 {top5.val:.2f} ({top5.avg:.2f})]".format( loss=base_losses, top1=base_top1, top5=base_top5 ) Astr = "Arch [Loss {loss.val:.3f} ({loss.avg:.3f}) Prec@1 {top1.val:.2f} ({top1.avg:.2f}) Prec@5 {top5.val:.2f} ({top5.avg:.2f})]".format( loss=arch_losses, top1=arch_top1, top5=arch_top5 ) logger.log(Sstr + " " + Tstr + " " + Wstr + " " + Astr) return base_losses.avg, base_top1.avg, base_top5.avg def valid_func(xloader, network, criterion): data_time, batch_time = AverageMeter(), AverageMeter() arch_losses, arch_top1, arch_top5 = AverageMeter(), AverageMeter(), AverageMeter() network.eval() end = time.time() with torch.no_grad(): for step, (arch_inputs, arch_targets) in enumerate(xloader): arch_targets = arch_targets.cuda(non_blocking=True) # measure data loading time data_time.update(time.time() - end) # prediction _, logits = network(arch_inputs) arch_loss = criterion(logits, arch_targets) # record arch_prec1, arch_prec5 = obtain_accuracy( logits.data, arch_targets.data, topk=(1, 5) ) arch_losses.update(arch_loss.item(), arch_inputs.size(0)) arch_top1.update(arch_prec1.item(), arch_inputs.size(0)) arch_top5.update(arch_prec5.item(), arch_inputs.size(0)) # measure elapsed time batch_time.update(time.time() - end) end = time.time() return arch_losses.avg, arch_top1.avg, arch_top5.avg def main(xargs): assert torch.cuda.is_available(), "CUDA is not available." torch.backends.cudnn.enabled = True torch.backends.cudnn.benchmark = False torch.backends.cudnn.deterministic = True torch.set_num_threads(xargs.workers) prepare_seed(xargs.rand_seed) logger = prepare_logger(args) train_data, valid_data, xshape, class_num = get_datasets( xargs.dataset, xargs.data_path, -1 ) # config_path = 'configs/nas-benchmark/algos/DARTS.config' config = load_config( xargs.config_path, {"class_num": class_num, "xshape": xshape}, logger ) search_loader, _, valid_loader = get_nas_search_loaders( train_data, valid_data, xargs.dataset, "configs/nas-benchmark/", config.batch_size, xargs.workers, ) logger.log( "||||||| {:10s} ||||||| Search-Loader-Num={:}, Valid-Loader-Num={:}, batch size={:}".format( xargs.dataset, len(search_loader), len(valid_loader), config.batch_size ) ) logger.log("||||||| {:10s} ||||||| Config={:}".format(xargs.dataset, config)) search_space = get_search_spaces("cell", xargs.search_space_name) if xargs.model_config is None: model_config = dict2config( { "name": "DARTS-V1", "C": xargs.channel, "N": xargs.num_cells, "max_nodes": xargs.max_nodes, "num_classes": class_num, "space": search_space, "affine": False, "track_running_stats": bool(xargs.track_running_stats), }, None, ) else: model_config = load_config( xargs.model_config, { "num_classes": class_num, "space": search_space, "affine": False, "track_running_stats": bool(xargs.track_running_stats), }, None, ) search_model = get_cell_based_tiny_net(model_config) logger.log("search-model :\n{:}".format(search_model)) w_optimizer, w_scheduler, criterion = get_optim_scheduler( search_model.get_weights(), config ) a_optimizer = torch.optim.Adam( search_model.get_alphas(), lr=xargs.arch_learning_rate, betas=(0.5, 0.999), weight_decay=xargs.arch_weight_decay, ) logger.log("w-optimizer : {:}".format(w_optimizer)) logger.log("a-optimizer : {:}".format(a_optimizer)) logger.log("w-scheduler : {:}".format(w_scheduler)) logger.log("criterion : {:}".format(criterion)) flop, param = get_model_infos(search_model, xshape) # logger.log('{:}'.format(search_model)) logger.log("FLOP = {:.2f} M, Params = {:.2f} MB".format(flop, param)) if xargs.arch_nas_dataset is None: api = None else: api = API(xargs.arch_nas_dataset) logger.log("{:} create API = {:} done".format(time_string(), api)) last_info, model_base_path, model_best_path = ( logger.path("info"), logger.path("model"), logger.path("best"), ) network, criterion = torch.nn.DataParallel(search_model).cuda(), criterion.cuda() if last_info.exists(): # automatically resume from previous checkpoint logger.log( "=> loading checkpoint of the last-info '{:}' start".format(last_info) ) last_info = torch.load(last_info) start_epoch = last_info["epoch"] checkpoint = torch.load(last_info["last_checkpoint"]) genotypes = checkpoint["genotypes"] valid_accuracies = checkpoint["valid_accuracies"] search_model.load_state_dict(checkpoint["search_model"]) w_scheduler.load_state_dict(checkpoint["w_scheduler"]) w_optimizer.load_state_dict(checkpoint["w_optimizer"]) a_optimizer.load_state_dict(checkpoint["a_optimizer"]) logger.log( "=> loading checkpoint of the last-info '{:}' start with {:}-th epoch.".format( last_info, start_epoch ) ) else: logger.log("=> do not find the last-info file : {:}".format(last_info)) start_epoch, valid_accuracies, genotypes = ( 0, {"best": -1}, {-1: search_model.genotype()}, ) # start training start_time, search_time, epoch_time, total_epoch = ( time.time(), AverageMeter(), AverageMeter(), config.epochs + config.warmup, ) for epoch in range(start_epoch, total_epoch): w_scheduler.update(epoch, 0.0) need_time = "Time Left: {:}".format( convert_secs2time(epoch_time.val * (total_epoch - epoch), True) ) epoch_str = "{:03d}-{:03d}".format(epoch, total_epoch) logger.log( "\n[Search the {:}-th epoch] {:}, LR={:}".format( epoch_str, need_time, min(w_scheduler.get_lr()) ) ) search_w_loss, search_w_top1, search_w_top5 = search_func( search_loader, network, criterion, w_scheduler, w_optimizer, a_optimizer, epoch_str, xargs.print_freq, logger, xargs.gradient_clip, ) search_time.update(time.time() - start_time) logger.log( "[{:}] searching : loss={:.2f}, accuracy@1={:.2f}%, accuracy@5={:.2f}%, time-cost={:.1f} s".format( epoch_str, search_w_loss, search_w_top1, search_w_top5, search_time.sum ) ) valid_a_loss, valid_a_top1, valid_a_top5 = valid_func( valid_loader, network, criterion ) logger.log( "[{:}] evaluate : loss={:.2f}, accuracy@1={:.2f}%, accuracy@5={:.2f}%".format( epoch_str, valid_a_loss, valid_a_top1, valid_a_top5 ) ) # check the best accuracy valid_accuracies[epoch] = valid_a_top1 if valid_a_top1 > valid_accuracies["best"]: valid_accuracies["best"] = valid_a_top1 genotypes["best"] = search_model.genotype() find_best = True else: find_best = False genotypes[epoch] = search_model.genotype() logger.log( "<<<--->>> The {:}-th epoch : {:}".format(epoch_str, genotypes[epoch]) ) # save checkpoint save_path = save_checkpoint( { "epoch": epoch + 1, "args": deepcopy(xargs), "search_model": search_model.state_dict(), "w_optimizer": w_optimizer.state_dict(), "a_optimizer": a_optimizer.state_dict(), "w_scheduler": w_scheduler.state_dict(), "genotypes": genotypes, "valid_accuracies": valid_accuracies, }, model_base_path, logger, ) last_info = save_checkpoint( { "epoch": epoch + 1, "args": deepcopy(args), "last_checkpoint": save_path, }, logger.path("info"), logger, ) if find_best: logger.log( "<<<--->>> The {:}-th epoch : find the highest validation accuracy : {:.2f}%.".format( epoch_str, valid_a_top1 ) ) copy_checkpoint(model_base_path, model_best_path, logger) with torch.no_grad(): # logger.log('arch-parameters :\n{:}'.format( nn.functional.softmax(search_model.arch_parameters, dim=-1).cpu() )) logger.log("{:}".format(search_model.show_alphas())) if api is not None: logger.log("{:}".format(api.query_by_arch(genotypes[epoch], "200"))) # measure elapsed time epoch_time.update(time.time() - start_time) start_time = time.time() logger.log("\n" + "-" * 100) logger.log( "DARTS-V1 : run {:} epochs, cost {:.1f} s, last-geno is {:}.".format( total_epoch, search_time.sum, genotypes[total_epoch - 1] ) ) if api is not None: logger.log("{:}".format(api.query_by_arch(genotypes[total_epoch - 1], "200"))) logger.close() if __name__ == "__main__": parser = argparse.ArgumentParser("DARTS first order") parser.add_argument("--data_path", type=str, help="Path to dataset") parser.add_argument( "--dataset", type=str, choices=["cifar10", "cifar100", "ImageNet16-120"], help="Choose between Cifar10/100 and ImageNet-16.", ) # channels and number-of-cells parser.add_argument("--search_space_name", type=str, help="The search space name.") parser.add_argument("--max_nodes", type=int, help="The maximum number of nodes.") parser.add_argument("--channel", type=int, help="The number of channels.") parser.add_argument( "--num_cells", type=int, help="The number of cells in one stage." ) parser.add_argument( "--track_running_stats", type=int, choices=[0, 1], help="Whether use track_running_stats or not in the BN layer.", ) parser.add_argument("--config_path", type=str, help="The config path.") parser.add_argument( "--model_config", type=str, help="The path of the model configuration. When this arg is set, it will cover max_nodes / channels / num_cells.", ) parser.add_argument("--gradient_clip", type=float, default=5, help="") # architecture leraning rate parser.add_argument( "--arch_learning_rate", type=float, default=3e-4, help="learning rate for arch encoding", ) parser.add_argument( "--arch_weight_decay", type=float, default=1e-3, help="weight decay for arch encoding", ) # log parser.add_argument( "--workers", type=int, default=2, help="number of data loading workers (default: 2)", ) parser.add_argument( "--save_dir", type=str, help="Folder to save checkpoints and log." ) parser.add_argument( "--arch_nas_dataset", type=str, help="The path to load the architecture dataset (nas-benchmark).", ) parser.add_argument("--print_freq", type=int, help="print frequency (default: 200)") parser.add_argument("--rand_seed", type=int, help="manual seed") args = parser.parse_args() if args.rand_seed is None or args.rand_seed < 0: args.rand_seed = random.randint(1, 100000) main(args)
[ "280835372@qq.com" ]
280835372@qq.com
158c8395e7b37a739bbe7438d2a3fb3853747fb2
0b20f4ce14b9ff77c84cedbecbaa29831335920d
/tests/cloudformation/file_formats/test_yaml.py
76149f86216a57acc3de965d65a22daae34bad5a
[ "Apache-2.0" ]
permissive
sergesec488/checkov
219c1b3864ab4f70b39a4cd79b041e98f3145364
56008e1c531b3626f14716067731be6e673040bc
refs/heads/master
2023-04-10T12:26:49.749864
2021-02-26T18:36:52
2021-02-26T18:40:58
342,883,133
0
1
Apache-2.0
2023-03-30T13:31:25
2021-02-27T15:01:08
null
UTF-8
Python
false
false
681
py
import os import unittest from checkov.cloudformation.runner import Runner from checkov.runner_filter import RunnerFilter class TestYamlFileFormat(unittest.TestCase): def test_summary(self): runner = Runner() current_dir = os.path.dirname(os.path.realpath(__file__)) test_files_dir = current_dir + "/yaml" report = runner.run(root_folder=test_files_dir) summary = report.get_summary() self.assertEqual(summary['passed'], 1) self.assertEqual(summary['failed'], 0) self.assertEqual(summary['skipped'], 0) self.assertEqual(summary['parsing_errors'], 0) if __name__ == '__main__': unittest.main()
[ "noreply@github.com" ]
noreply@github.com
cb73af820437abcd75867aff99c64c43744110ef
c2c1fffabcdbc81cedfd039a1090000ff70f6968
/Proxy/ProxyGetter.py
2243f29bd3d3d11d82bf42cdbb1e589307628c55
[]
no_license
nightstalkerx/ProxyPool
c92eafa9badc32bf085d44f9493e3b2a805bc602
3982b293f6091fae36f39cc0683ec47ebf2020d1
refs/heads/master
2020-04-21T23:25:11.544582
2018-09-05T11:01:16
2018-09-05T11:01:16
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,737
py
import sys sys.path.append('../') from Util.UtilFunction import get_html_tree from Util.WebRequest import WebRequest class ProxyGetter(): """ 代理获取类 """ def __init__(self): pass @staticmethod def get_proxy_one(): """ 采集:http://www.ip181.com/ :return: """ url = 'http://www.ip181.com/' html = get_html_tree(url) nodies = html.xpath('//tr')[1:] for node in nodies: ip_port = node.xpath('.//td/text()')[0:2] yield ':'.join(ip_port) @staticmethod def get_proxy_two(): """ 采集:http://www.xdaili.cn/freeproxy :return: """ url = 'http://www.xdaili.cn/ipagent/freeip/getFreeIps?page=1&rows=10' request = WebRequest() res = request.get(url).json() for row in res['RESULT']['rows']: yield '{}:{}'.format(row['ip'], row['port']) @staticmethod def get_proxy_three(): """ 采集:https://www.kuaidaili.com :return: """ url = 'https://www.kuaidaili.com/free/inha/{}/' for i in range(1, 5): html = get_html_tree(url.format(i)) nodies = html.xpath('//tr') for node in nodies: ip_port = node.xpath('.//td/text()')[0:2] yield ':'.join(ip_port) @staticmethod def get_proxy_forth(): url = 'http://www.mogumiao.com/proxy/free/listFreeIp' request = WebRequest() res = request.get(url).json() for row in res['msg']: yield '{}:{}'.format(row['ip'], row['port']) if __name__ == '__main__': for i in ProxyGetter.get_proxy_fifth(): print(i)
[ "610426459@qq.com" ]
610426459@qq.com
9e2de46d09baa08db907b3cdb46d3b08b3da73b1
87b906a183d6cf7f041356c8ff5aa2ee59d03a09
/Day 3/Ticket for photo.py
5e9236eaea7dbf2d0f4cf341142f621033e8885c
[]
no_license
amitshrestha-15/100daysof-code
558635f50c26568fb6186bf648e2b783bbc7d7c1
16969e8562ed49a88fbe010e02e27108ff5470c7
refs/heads/master
2023-05-21T08:07:50.820307
2021-06-04T11:53:23
2021-06-04T11:53:23
373,150,913
0
0
null
null
null
null
UTF-8
Python
false
false
711
py
print ("Welcome to roller coaster!") height = int(input("What is your height in cm?")) bill = 0 if height >= 120 : age = int(input("What is your age?")) if age < 12: bill = 5 print("Child ticket price is $5.") elif age <= 18: bill = 7 print("Youth ticket price is $7.") elif age > 18 and age< 45: bill = 12 print("Adult ticket price is $12.") elif age>= 45 and age <= 55: print("Your ride is free. But You should pay for photo.") want_photo = input("Do you want a photo taken? Y or N. ") if want_photo == "Y": bill += 3 print(f"Your ticket price is ${bill}") else: print("Sorry! You can't ride roller coaster.")
[ "amitstha1234@gmail.com" ]
amitstha1234@gmail.com
7556cf1918258fe40c2aabedcaf970d5535e3a6b
7c926109cda8e59cef6abf097774c1d64c7f77af
/app.py
18bb8ed0dbd02ac640d073f871277c93e192c80d
[]
no_license
zahraEskandari/corona_dashboard
f27ee1091a24cf622d1f631712036387760e6327
4c881800369a9c3be390409e4d9d146fbd5858c2
refs/heads/master
2021-05-20T04:15:44.021841
2020-04-01T13:24:15
2020-04-01T13:24:15
252,181,036
0
0
null
null
null
null
UTF-8
Python
false
false
29,609
py
#from flask import Flask import flask import dash import dash_core_components as dcc import dash_html_components as html import pandas as pd import plotly.graph_objs as go import random import math import dash_table as dst from dash.dependencies import Input, Output import datetime import plotly # storing and anaysis import numpy as np # visualization #import matplotlib.pyplot as plt #import seaborn as sns import plotly.express as px import folium #from time import strftime from flask_caching import Cache import json data_path = 'path to .. \\data\\corona_data.csv' external_scripts = [ {'src': 'my view counter script'} ] server = flask.Flask(__name__) app = dash.Dash(__name__, server=server, external_scripts=external_scripts) # show data def serve_layout(): # importing datasets full_table = pd.read_csv(data_path , parse_dates=['Date']) # unifying names full_table['Country/Region'] = full_table['Country/Region'].replace('Mainland China', 'China') full_table['Country/Region'] = full_table['Country/Region'].replace('Iran (Islamic Republic of)', 'Iran') full_table['Country/Region'] = full_table['Country/Region'].replace('Republic of Korea', 'South Korea') full_table['Country/Region'] = full_table['Country/Region'].replace('Korea, South', 'South Korea') # filling missing values with NA full_table[['Province/State']] = full_table[['Province/State']].fillna('NA') Iran = full_table[full_table['Country/Region']=='Iran'] full_latest = full_table[full_table['Date'] == max(full_table['Date'])].reset_index() # last 2 days temp = full_table.groupby('Date')['Confirmed', 'Deaths', 'Recovered'].sum() temp = temp.reset_index() temp = temp.sort_values('Date', ascending=False) today = temp.iloc[0]["Date"] t = str(today)#today.strftime("%d-%b-%Y") today = t[0:t.find('T')] total_cases_today = temp.iloc[0]["Confirmed"] death_cases_today = temp.iloc[0]["Deaths"] recovered_cases_today = temp.iloc[0]["Recovered"] print(temp.head()) yesterday = temp.iloc[1]["Date"] total_cases_yesterday = temp.iloc[1]["Confirmed"] death_cases_yesterday = temp.iloc[1]["Deaths"] recovered_cases_yesterday = temp.iloc[1]["Recovered"] temp_Iran = Iran.groupby('Date')['Confirmed', 'Deaths', 'Recovered'].sum() temp_Iran = temp_Iran.reset_index() temp_Iran = temp_Iran.sort_values('Date', ascending=False) iran_today = temp_Iran.iloc[0]["Date"] iran_total_cases_today = temp_Iran.iloc[0]["Confirmed"] iran_death_cases_today = temp_Iran.iloc[0]["Deaths"] iran_recovered_cases_today = temp_Iran.iloc[0]["Recovered"] iran_yesterday = temp_Iran.iloc[1]["Date"] iran_total_cases_yesterday = temp_Iran.iloc[1]["Confirmed"] iran_death_cases_yesterday = temp_Iran.iloc[1]["Deaths"] iran_recovered_cases_yesterday = temp_Iran.iloc[1]["Recovered"] full_latest = full_table[full_table['Date'] == max(full_table['Date'])].reset_index() full_latest_grouped = full_latest.groupby('Country/Region')['Confirmed', 'Deaths', 'Recovered'].sum().reset_index() full_latest_grouped_confirmed = full_latest_grouped[['Country/Region', 'Confirmed']] result = full_latest_grouped_confirmed.nlargest(8, columns='Confirmed') print(int(full_latest_grouped_confirmed.loc[full_latest_grouped_confirmed['Country/Region']=='Iran'] ["Confirmed"])) if 'Iran' not in result['Country/Region'].values : result.loc[len(result)] = ['Iran', int(full_latest_grouped_confirmed.loc[full_latest_grouped_confirmed['Country/Region']=='Iran'] ["Confirmed"])] result.loc[len(result)] = ['Other Countries', full_latest_grouped_confirmed.loc[~full_latest_grouped_confirmed['Country/Region'].isin(result['Country/Region']), 'Confirmed'].sum()] #full_latest_grouped_confirmed result temp_full = full_table.groupby(['Country/Region', 'Date'])['Confirmed', 'Deaths', 'Recovered'].sum() temp_full = temp_full.reset_index() #temp_full['Country/Region'].isin(result['Country/Region']) temp = temp_full.loc[temp_full['Country/Region'].isin(result['Country/Region']) ] temp2 = temp_full.loc[~temp_full['Country/Region'].isin(result['Country/Region']) ].groupby(['Date'])['Confirmed', 'Deaths', 'Recovered'].sum() temp2 = temp2.reset_index() temp2['Country/Region'] = 'Other Countries' temp = temp.append(temp2, ignore_index=True) temp fig1 = px.bar(temp, x="Date", y="Confirmed", color='Country/Region', orientation='v',width= 600 , height=600, title='مجموع موارد تایید شده در دنیا', color_discrete_sequence = px.colors.cyclical.HSV) fig1.update_layout(legend_orientation='h') #fig.show() fig2 = px.bar(temp, x="Date", y="Deaths", color='Country/Region', orientation='v', width= 600 , height=600, title='مجموع موارد فوت شده در دنیا', color_discrete_sequence = px.colors.cyclical.HSV) fig2.update_layout(legend_orientation='h') #fig.show() fig3= px.line(temp, x='Date', y='Confirmed', color='Country/Region', width= 600 , height=600, title=' موارد تایید شده به تفکیک کشور', color_discrete_sequence = px.colors.cyclical.HSV ) fig3.update_layout(legend_orientation='h')#fig.show() fig4= px.line(temp, x='Date', y='Deaths', color='Country/Region' , width= 600 , height=600, title=' موارد فوت شده به تفکیک کشور', color_discrete_sequence = px.colors.cyclical.HSV ) fig4.update_layout(legend_orientation='h')#fig.show() gdf = gdf = full_table.groupby(['Date', 'Country/Region'])['Confirmed', 'Deaths', 'Recovered'].max() gdf = gdf.reset_index() temp_iran = gdf[gdf['Country/Region']=='Iran'].groupby('Date').sum().reset_index() temp_iran = temp_iran.melt(id_vars='Date', value_vars=['Confirmed', 'Deaths', 'Recovered'], var_name='Case', value_name='Count') fig5 = px.bar(temp_iran, x="Date", y="Count", color='Case', facet_col="Case", title='مجموع موارد تایید شده، فوت شده و بهبود یافته ایران' , width=1000) temp_iran2 = Iran.groupby('Date')['Confirmed', 'Deaths', 'Recovered'].sum().diff() print(temp_iran2) temp_iran2 = temp_iran2.reset_index() temp_iran2 = temp_iran2.melt(id_vars="Date", value_vars=['Confirmed', 'Deaths', 'Recovered']) fig9 = px.bar(temp_iran2, x="Date", y="value", color='variable', title='تعداد موارد تایید شده، فوت شده و بهبود یافته ایران در هر روز' , width=1000) fig9.update_layout(barmode='group') temp['Mortality Rate'] = round(1.0 * temp['Deaths']/ temp['Confirmed'], 3)*100 temp['Recovery Rate'] = round(1.0 * temp['Recovered']/ temp['Confirmed'], 3)*100 #fig5.show() fig6 = px.line(temp, x="Date", y='Mortality Rate', color='Country/Region', facet_col='Country/Region',hover_name='Country/Region' , facet_col_wrap = 3 , render_mode = 'webgl' , title='نرخ موارد منجر به مرگ نسبت به موارد تایید شده در طول زمان' , width=1000 , height=700) fig6.for_each_annotation(lambda a: a.update(text=a.text.split("=")[-1])) fig7 = px.line(temp, x="Date", y='Recovery Rate', color='Country/Region', facet_col='Country/Region',hover_name='Country/Region' , facet_col_wrap = 3 , render_mode = 'webgl' , title='نرخ موارد بهبود یافته نسبت به موارد تایید شده در طول زمان' , width=1000 , height=700) fig7.for_each_annotation(lambda a: a.update(text=a.text.split("=")[-1])) fig8 = px.line(temp, x="Date", y='Confirmed', color='Country/Region', facet_col='Country/Region',hover_name='Country/Region' , facet_col_wrap = 3 , render_mode = 'auto' , title='تعداد موارد تایید شده در کشورهای مختلف' , width=1000 , height=700) fig8.for_each_annotation(lambda a: a.update(text=a.text.split("=")[-1])) colors = { 'background': '#222222', 'text': '#7FDBFF' } d1 = html.Div( id = 'd1' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px' } , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد تایید شده در دنیا' )) ,html.Div( id ='confirmedCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format(int(total_cases_today)) ) # ,html.Div( className = 'content' , children = "{:,d}".format(int(total_cases_today - total_cases_yesterday)) + ' مورد بیشتر از روز قبل' ) #"{:,d}".format(total_cases_today - total_cases_yesterday) + ' مورد بیشتر از روز قبل' ] ) , html.Div( className = 'metadata' , children='تاریخ به روز رسانی:' + str(today) ) #'تاریخ به روز رسانی:' + today.strftime("%b %d %Y") ] ) d2 = html.Div( id = 'd2' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px'} , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد فوت شده در دنیا' )) ,html.Div( id ='DeathsCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format(int(death_cases_today)) ) #str(total_cases_today) ,html.Div( className = 'content' , children = "{:,d}".format(int(death_cases_today - death_cases_yesterday) ) + ' مورد بیشتر از روز قبل' ) ] ) , html.Div( className = 'metadata' , children= 'تاریخ به روز رسانی:' + str(today) ) ] ) d3 = html.Div( id = 'd3' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px'} , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد بهبود یافته در دنیا' )) ,html.Div( id ='RecoveredCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format(int(recovered_cases_today)) ) #str(total_cases_today) ,html.Div( className = 'content' , children = "{:,d}".format(int(recovered_cases_today - recovered_cases_yesterday)) + ' مورد بیشتر از دیروز' ) ] ) , html.Div( className = 'metadata' , children= 'تاریخ به روز رسانی:' + str(today) )#today.strftime("%b %d %Y") ] ) d1_iran = html.Div( id = 'd1_iran' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px' } , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد تایید شده در ایران' )) ,html.Div( id ='iran_confirmedCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format( int(iran_total_cases_today)) ) #str(total_cases_today) ,html.Div( className = 'content' , children = "{:,d}".format( int(iran_total_cases_today - iran_total_cases_yesterday)) + ' مورد بیشتر از روز قبل' ) ] ) , html.Div( className = 'metadata' , children= 'تاریخ به روز رسانی:' + str(today) ) ] ) d2_iran = html.Div( id = 'd2_iran' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px'} , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد فوت شده در ایران' )) ,html.Div( id ='iran_DeathsCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format( int(iran_death_cases_today)) ) #str(total_cases_today) ,html.Div( className = 'content' , children = "{:,d}".format(int(iran_death_cases_today - iran_death_cases_yesterday) ) + ' مورد بیشتر از روز قبل' ) ] ) , html.Div( className = 'metadata' , children= 'تاریخ به روز رسانی:' + str(today) ) ] ) d3_iran = html.Div( id = 'd3_iran' , className = 'class list emergency' , style={'color':'white' , 'border':'solid 1px white' , 'direction':'rtl' , 'padding':'20px' , 'margin':'8px'} , children=[ html.Div( className = 'content' , children=[ html.Div( className = 'content' , children = html.Span( style={'font-size': '24px' } ,children='موارد بهبود یافته در ایران' )) ,html.Div( id ='iran_RecoveredCases_count' , style= {'color':'red' , 'font-size': '48px' } , className = 'heading1' , children = "{:,d}".format(int(iran_recovered_cases_today)) ) #str(total_cases_today) ,html.Div( className = 'content' , children = "{:,d}".format(int(iran_recovered_cases_today - iran_recovered_cases_yesterday)) + ' مورد بیشتر از روز قبل' ) ] ) , html.Div( className = 'metadata' , children= 'تاریخ به روز رسانی:' + str(today) ) ] ) confirmed_cases_fig_div = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='life-exp-vs-gdp6', figure = fig1) ) ] ) confirmed_cases_fig_div_2 = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='confirmed_cases_fig_div_2', figure = fig3) ) ] ) deaths_cases_fig_div = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='life-exp-vs-gdp7', figure = fig2) ) ] ) deaths_cases_fig_div_2 = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='deaths_cases_fig_div_2', figure = fig4) ) ] ) iran_status = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='iran_status', figure = fig5) ) ] ) iran_daily_status = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='iran_daily_status', figure = fig9) ) ] ) mortality_rate = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='mortality_rate', figure = fig6) ) ] ) recovery_rate = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='reconvery_rate', figure = fig7) ) ] ) confirmed_cases_faceted = html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} , children=[ html.Div( style={'display': 'flex' , 'flex-direction': 'row' , 'margin-left' :'10px', 'margin-right' :'10px'} , children = dcc.Graph( id='confirmed_cases_faceted', figure = fig8) ) ] ) date_string = f'{datetime.datetime.now():%Y-%m-%d %H:%M:%S%z}' return html.Div( style={ 'display': 'flex' , 'flex-direction': 'column' , 'justify-content': 'center' , 'backgroundColor': colors['background']} ,children=[ html.Div( children=[ html.H1(style={'color':'white' , 'text-align':'center'} , children='وضعیت انتشار کرونا در ایران و دنیا') , html.H5(style={'color':'white' , 'text-align':'center'} , children='Loaded at: ' +date_string) ] ) , html.Div( id="global_stats" , children = [ d1 , d2 , d3] , style= { 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'20px' , 'backgroundColor': colors['background'] , 'direction':'rtl'} ) , html.Div( id="iran_stats" , children = [ d1_iran , d2_iran , d3_iran] , style= { 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'30px' , 'backgroundColor': colors['background'] , 'direction':'rtl'} ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children=[ confirmed_cases_fig_div , deaths_cases_fig_div ] ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children=[ confirmed_cases_fig_div_2 , deaths_cases_fig_div_2 ] ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = iran_status ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = iran_daily_status ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = confirmed_cases_faceted ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = mortality_rate ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = recovery_rate ) #, html.A(href="http://www.webgozar.com/counter/stats.aspx?code=3745510" , target="_blank" , children='&#1570;&#1605;&#1575;&#1585;') , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = [ html.Span( style= {'color':'white'} , children=' این داشبورد بر اساس داده‌های آماری دانشگاه جان هاپکینز آمریکا تهیه شده است. از طریق لینک زیر می توانید به این داده ها دسترسی پیدا کنید. ') ] ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = [ html.A(style= {'color':'white'} , href="https://github.com/CSSEGISandData/COVID-19" , target="_blank" , children='https://github.com/CSSEGISandData/COVID-19') ,html.Br() ] ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = [ html.Span( style= {'color':'white'} , children='برای ارتباط با من می توانید روی وبلاگ دیتا اینسایتز به آدرس زیر پیام بگذارید : ') #, html.A(style= {'color':'white'} , href="https://datainsights.blogsky.com/" , target="_blank" , children=' دیتا اینسایتز ') ] ) , html.Div( style={ 'display': 'flex' , 'flex-direction': 'row' , 'justify-content': 'center' , 'margin-bottom' :'10px' , 'backgroundColor': colors['background'] , 'direction' : 'rtl'} , children = [ html.A(style= {'color':'white'} , href="https://datainsights.blogsky.com/" , target="_blank" , children='https://datainsights.blogsky.com/ ') ] ) ]) app.layout = serve_layout app.scripts.append_script({"external_url": "http://www.webgozar.ir/c.aspx?Code=3570732&amp;t=counter"}) if __name__ == '__main__': #app.run_server(debug=True) app.run_server(debug=False , dev_tools_hot_reload=False)
[ "zahra.eskandari@gmail.com" ]
zahra.eskandari@gmail.com
f60987e55994a05e1fbf45fa4d8ded677baca05b
732374714ffe0e0f2c07a493a2ee71c9271fdce0
/mysite/settings.py
bcd771fb691401a56d55a3106a4ee650b115e261
[]
no_license
aaronahmid/mosunhomesrealtors
721fb20d671f1a58c64abc8bdf1209a5ab3236f1
561b56fd90179e163f0c861dae1d451cc1cfc662
refs/heads/main
2023-08-13T02:22:46.005517
2021-10-09T05:15:59
2021-10-09T05:15:59
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,269
py
""" Django settings for mysite project. Generated by 'django-admin startproject' using Django 3.2.7. For more information on this file, see https://docs.djangoproject.com/en/3.2/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/3.2/ref/settings/ """ import os import dj_database_url import django_heroku import cloudinary import cloudinary.uploader import cloudinary.api from pathlib import Path # Build paths inside the project like this: BASE_DIR / 'subdir'. BASE_DIR = Path(__file__).resolve().parent.parent # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'django-insecure--+3$m3fs+h3qdye&74^k@qadoro606d*%%qacpzw=&7g!ruu@l' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = False ALLOWED_HOSTS = ['127.0.0.1', '.herokuapp.com', 'www.mosunhomes-realtors.com', 'mosunhomes-realtors.com'] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'blog.apps.BlogConfig', 'cloudinary', 'cloudinary_storage', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'whitenoise.middleware.WhiteNoiseMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'mysite.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'mysite.wsgi.application' # Database # https://docs.djangoproject.com/en/3.2/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } # Password validation # https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/3.2/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'Africa/Lagos' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/3.2/howto/static-files/ PROJECT_ROOT = os.path.dirname(os.path.abspath(__file__)) STATIC_URL = '/static/' STATICFILES_DIRS = os.path.join(BASE_DIR, "blog/static"), STATIC_ROOT = os.path.join(BASE_DIR, 'staticfiles') MEDIA_URL = '/media/' MEDIA_ROOT = os.path.join(BASE_DIR, 'media') STATICFILES_STORAGE = 'whitenoise.storage.CompressedManifestStaticFilesStorage' cloudinary.config( cloud_name = "thormiwa", api_key = "584634363435482", api_secret = "XGzynridSBzxfDGpkyOMnHAHGrA" ) DEFAULT_FILE_STORAGE = 'cloudinary_storage.storage.MediaCloudinaryStorage' # Default primary key field type # https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field DEFAULT_AUTO_FIELD = 'django.db.models.BigAutoField' SECURE_PROXY_SSL_HEADER = ('HTTP_X_FORWARDED_PROTO', 'https') SECURE_SSL_REDIRECT = True # Activate Django-Heroku. django_heroku.settings(locals())
[ "thormiwa04@gmail.com" ]
thormiwa04@gmail.com
b29f5d7cbbd6924909c5ec897b40f7a01208f045
92115d37e1199a21367986cb97ddc64580432d34
/seq2seq/data_dial/create_dial_data.py
5f8bb9d4d1605857aa6028183a295eac6d35494e
[]
no_license
MasatoMiyoshi/tensorflow_tutorial
9575acb495ec16bb91fdbc8b9e22bee7b451ac0a
cc51632f818aa2e2c73400fb6348d108e41d66bf
refs/heads/master
2021-07-13T18:02:33.072835
2020-05-11T09:10:58
2020-05-11T09:11:27
129,721,097
0
0
null
null
null
null
UTF-8
Python
false
false
1,123
py
# -*- coding: utf-8 -*- from __future__ import division, print_function, unicode_literals, absolute_import import sys import os import codecs import re import glob import commands reload(sys) sys.setdefaultencoding('utf-8') dirname = "./json/init100" fh_u = codecs.open('src.txt', 'w', 'utf-8') fh_s = codecs.open('tgt.txt', 'w', 'utf-8') filelist = glob.glob(dirname + "/*.json") for filename in filelist: outputs = commands.getoutput(("python show_dial_fix.py " + filename)) lines = outputs.split("\n") has_system_uttr = False system_uttr = '' has_user_uttr = False user_uttr = '' for line in lines: if line.startswith("S:") and has_user_uttr: has_system_uttr = True system_uttr = re.sub(r'^S:', '', line) if line.startswith("U:"): has_user_uttr = True user_uttr = re.sub(r'^U:', '', line) if has_system_uttr and has_user_uttr: fh_u.write(user_uttr + "\n") fh_s.write(system_uttr + "\n") has_system_uttr = False has_user_uttr = False fh_u.close() fh_s.close() ### EOF
[ "mirasi1h@gmail.com" ]
mirasi1h@gmail.com
016624cd89be6e7f3e25cc65d992a04808dc6ae2
9d008ba29c32180518e03a58e5b3cb946c8a1d6c
/app.py
fdfaa5b29ad43f850c5077d83e2b36ee18fab768
[]
no_license
MikeCarbone/Personal-website
9f5f25e031ca8be939fa26df89b44b1da51be871
16540283128cc7b711886dfdf2b99abd7d3c9af5
refs/heads/master
2021-09-02T06:10:44.005130
2017-12-30T23:19:00
2017-12-30T23:19:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
916
py
import tornado.ioloop import tornado.web import tornado.httpserver import os.path class MainHandler(tornado.web.RequestHandler): def get(self): self.render('main.html') class tempHandler(tornado.web.RequestHandler): def get(self): self.render('underConstruction.html') settings = dict( template_path = os.path.join(os.path.dirname(__file__), "templates"), static_path = os.path.join(os.path.dirname(__file__), "static"), debug = False ) handlers = [(r'/', MainHandler), (r'/underConstruction', tempHandler)] def app(): print('Server Running...') print('Press ctrl + c to close') application = tornado.web.Application(handlers, **settings) http_server = tornado.httpserver.HTTPServer(application) port = int(os.environ.get("PORT", 5000)) http_server.listen(port) #application.listen(8888) tornado.ioloop.IOLoop.instance().start() #Start the server at port n if __name__ == "__main__": app()
[ "mfcbone@gmail.com" ]
mfcbone@gmail.com
db053d106eb43e3318aeb6cca68e054a75650c83
77cc444d492545c322da1c7819e81a6542e3a6ff
/likes/models.py
b98a20354d3401cce7bab157417a960e8de39b09
[ "Apache-2.0" ]
permissive
PingLu8/django-twitter
3bd3a646c8d74179160f8fad11c1fa40018d59e7
ef5c47a9b2cd6796d67028946b2ed1c7b0af51be
refs/heads/main
2023-07-08T05:38:47.605500
2021-08-11T04:00:07
2021-08-11T04:00:07
361,040,505
0
0
Apache-2.0
2021-08-11T04:00:08
2021-04-24T00:58:36
Python
UTF-8
Python
false
false
1,320
py
from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.fields import GenericForeignKey from utils.memcached_helper import MemcachedHelper from django.db.models.signals import pre_delete, post_save from likes.listeners import incr_likes_count, decr_likes_count class Like(models.Model): object_id = models.PositiveIntegerField() # comment_id or tweet_id content_type = models.ForeignKey( ContentType, on_delete=models.SET_NULL, null=True, ) content_object = GenericForeignKey('content_type', 'object_id') user = models.ForeignKey(User, on_delete=models.SET_NULL, null=True ) created_at = models.DateTimeField(auto_now_add=True) class Meta: unique_together = ( ('user', 'content_type', 'object_id'), ) index_together = ( ( 'content_type', 'object_id', 'created_at'), ) def __str__(self): return f'{self.created_at} - {self.user} liked {self.content_type} {self.object_id}' @property def cached_user(self): return MemcachedHelper.get_object_through_cache(User, self.user_id) pre_delete.connect(decr_likes_count, sender=Like) post_save.connect(incr_likes_count, sender=Like)
[ "noreply@github.com" ]
noreply@github.com
b4cebd6904d477cd8224278ad3c87bbe2000ae9e
ccbfc7818c0b75929a1dfae41dc061d5e0b78519
/aliyun-openapi-python-sdk-master/aliyun-python-sdk-vpc/aliyunsdkvpc/request/v20160428/CreateRouterInterfaceRequest.py
f3794b0030c799277bdbb14c640f9f31c41bee1c
[ "Apache-2.0" ]
permissive
P79N6A/dysms_python
44b634ffb2856b81d5f79f65889bfd5232a9b546
f44877b35817e103eed469a637813efffa1be3e4
refs/heads/master
2020-04-28T15:25:00.368913
2019-03-13T07:52:34
2019-03-13T07:52:34
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,414
py
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. from aliyunsdkcore.request import RpcRequest class CreateRouterInterfaceRequest(RpcRequest): def __init__(self): RpcRequest.__init__(self, 'Vpc', '2016-04-28', 'CreateRouterInterface','vpc') def get_AccessPointId(self): return self.get_query_params().get('AccessPointId') def set_AccessPointId(self,AccessPointId): self.add_query_param('AccessPointId',AccessPointId) def get_OppositeRouterId(self): return self.get_query_params().get('OppositeRouterId') def set_OppositeRouterId(self,OppositeRouterId): self.add_query_param('OppositeRouterId',OppositeRouterId) def get_OppositeAccessPointId(self): return self.get_query_params().get('OppositeAccessPointId') def set_OppositeAccessPointId(self,OppositeAccessPointId): self.add_query_param('OppositeAccessPointId',OppositeAccessPointId) def get_ResourceOwnerId(self): return self.get_query_params().get('ResourceOwnerId') def set_ResourceOwnerId(self,ResourceOwnerId): self.add_query_param('ResourceOwnerId',ResourceOwnerId) def get_Role(self): return self.get_query_params().get('Role') def set_Role(self,Role): self.add_query_param('Role',Role) def get_ClientToken(self): return self.get_query_params().get('ClientToken') def set_ClientToken(self,ClientToken): self.add_query_param('ClientToken',ClientToken) def get_HealthCheckTargetIp(self): return self.get_query_params().get('HealthCheckTargetIp') def set_HealthCheckTargetIp(self,HealthCheckTargetIp): self.add_query_param('HealthCheckTargetIp',HealthCheckTargetIp) def get_Description(self): return self.get_query_params().get('Description') def set_Description(self,Description): self.add_query_param('Description',Description) def get_Spec(self): return self.get_query_params().get('Spec') def set_Spec(self,Spec): self.add_query_param('Spec',Spec) def get_OppositeInterfaceId(self): return self.get_query_params().get('OppositeInterfaceId') def set_OppositeInterfaceId(self,OppositeInterfaceId): self.add_query_param('OppositeInterfaceId',OppositeInterfaceId) def get_InstanceChargeType(self): return self.get_query_params().get('InstanceChargeType') def set_InstanceChargeType(self,InstanceChargeType): self.add_query_param('InstanceChargeType',InstanceChargeType) def get_Period(self): return self.get_query_params().get('Period') def set_Period(self,Period): self.add_query_param('Period',Period) def get_AutoPay(self): return self.get_query_params().get('AutoPay') def set_AutoPay(self,AutoPay): self.add_query_param('AutoPay',AutoPay) def get_ResourceOwnerAccount(self): return self.get_query_params().get('ResourceOwnerAccount') def set_ResourceOwnerAccount(self,ResourceOwnerAccount): self.add_query_param('ResourceOwnerAccount',ResourceOwnerAccount) def get_OppositeRegionId(self): return self.get_query_params().get('OppositeRegionId') def set_OppositeRegionId(self,OppositeRegionId): self.add_query_param('OppositeRegionId',OppositeRegionId) def get_OwnerAccount(self): return self.get_query_params().get('OwnerAccount') def set_OwnerAccount(self,OwnerAccount): self.add_query_param('OwnerAccount',OwnerAccount) def get_OwnerId(self): return self.get_query_params().get('OwnerId') def set_OwnerId(self,OwnerId): self.add_query_param('OwnerId',OwnerId) def get_OppositeInterfaceOwnerId(self): return self.get_query_params().get('OppositeInterfaceOwnerId') def set_OppositeInterfaceOwnerId(self,OppositeInterfaceOwnerId): self.add_query_param('OppositeInterfaceOwnerId',OppositeInterfaceOwnerId) def get_RouterType(self): return self.get_query_params().get('RouterType') def set_RouterType(self,RouterType): self.add_query_param('RouterType',RouterType) def get_HealthCheckSourceIp(self): return self.get_query_params().get('HealthCheckSourceIp') def set_HealthCheckSourceIp(self,HealthCheckSourceIp): self.add_query_param('HealthCheckSourceIp',HealthCheckSourceIp) def get_RouterId(self): return self.get_query_params().get('RouterId') def set_RouterId(self,RouterId): self.add_query_param('RouterId',RouterId) def get_OppositeRouterType(self): return self.get_query_params().get('OppositeRouterType') def set_OppositeRouterType(self,OppositeRouterType): self.add_query_param('OppositeRouterType',OppositeRouterType) def get_Name(self): return self.get_query_params().get('Name') def set_Name(self,Name): self.add_query_param('Name',Name) def get_PricingCycle(self): return self.get_query_params().get('PricingCycle') def set_PricingCycle(self,PricingCycle): self.add_query_param('PricingCycle',PricingCycle)
[ "1478458905@qq.com" ]
1478458905@qq.com
93b57b5d8ab7beae315d919322890e775a1998e9
53fab060fa262e5d5026e0807d93c75fb81e67b9
/backup/user_188/ch78_2019_04_04_19_41_08_100209.py
61335cb39d3ea9f3ad223e32472973ee949e080e
[]
no_license
gabriellaec/desoft-analise-exercicios
b77c6999424c5ce7e44086a12589a0ad43d6adca
01940ab0897aa6005764fc220b900e4d6161d36b
refs/heads/main
2023-01-31T17:19:42.050628
2020-12-16T05:21:31
2020-12-16T05:21:31
306,735,108
0
0
null
null
null
null
UTF-8
Python
false
false
1,203
py
from math import sqrt def calcula_tempo(atletas): tempo_de_conclusao = {} for nome in atletas: tempo_atleta = sqrt(200 / atletas[nome]) tempo_de_conclusao[nome] = tempo_atleta return tempo_de_conclusao def atleta_mais_rapido(dicionario): menor_tempo = 0 melhor_atleta = "" for nome in dicionario: if menor_tempo > dicionario[nome]: menor_tempo = dicionario[nome] melhor_atleta = nome return melhor_atleta def tempo_mais_curto(dicionario): menor_tempo = 0 for nome in dicionario: if menor_tempo > dicionario[nome]: menor_tempo = dicionario[nome] return menor_tempo nomes_aceleracoes_ateltas = {} sair = False while not sair: nome = input("Digite o nome do atleta: ") aceleracao = int(input("Digite a aceleracao do atleta: ")) if nome == "sair": sair = True else: nomes_aceleracoes_atletas[nome] = aceleracao nomes_tempos_atletas = calcula_tempo(nomes_aceleracoes_atletas) nome = atleta_mais_rapido(nomes_tempos_atletas) tempo = tempo_mais_curto(nomes_tempos_atletas) print('O vencedor é {0} com tempo de conclusão de {1} s'.format(nome, tempo))
[ "you@example.com" ]
you@example.com
40719b630af8856497ca0e697dd33dee816eeeb1
ac042704660f07263a9b7918c9d19e8027e2c01b
/qn 9.py
adc05564fc75215eb2d78a4b3c7011f3db70e44f
[]
no_license
Prashant414/python
23387f2d205ceb36f141e4b4529ff9c3e80d2679
f5ff2b280b4bf29df2723b9d1d16690e65aaf62f
refs/heads/main
2023-03-21T10:23:32.119726
2021-03-09T17:59:45
2021-03-09T17:59:45
null
0
0
null
null
null
null
UTF-8
Python
false
false
434
py
# Modify the above question to allow student to sit if he/she has medical cause. Ask user if he/she has medical cause or not ( 'Y' or 'N' ) anx=float(input("number of clases held")) x=float(input("number of clases held")) a=float(input("number of classes attend")) m=input("any medical issue type y or n") z=a/x*100 if z<=75: if m=='y': print("allowed") else: print("not allowed") else: print("allowed")
[ "prashantshyam09@gmail.com" ]
prashantshyam09@gmail.com
8effb973e9606af222d40da9faf37e6076c24cca
7dac8d38552a8f8eb401158ed3aabffaf8f23251
/gobotany/site/migrations/0008_alter_document_upload.py
c2b8ece6b53b86964306fb7bec83afeba21b88a8
[]
no_license
newfs/gobotany-app
3b9933a55a2b7ab5adcdd10aa371b104c1a0850b
9030d08b9a1b8bdb0f897c6e482c09ef78cc4d3d
refs/heads/master
2023-01-11T02:49:04.585248
2022-11-17T00:56:30
2022-11-17T00:56:30
3,584,135
10
6
null
2016-02-06T23:07:24
2012-02-29T17:49:06
JavaScript
UTF-8
Python
false
false
497
py
# Generated by Django 3.2.15 on 2022-11-11 00:36 from django.db import migrations, models import storages.backends.s3boto class Migration(migrations.Migration): dependencies = [ ('site', '0007_auto_20210602_2000'), ] operations = [ migrations.AlterField( model_name='document', name='upload', field=models.FileField(storage=storages.backends.s3boto.S3BotoStorage(bucket='newfs', location='docs'), upload_to=''), ), ]
[ "jnga@users.noreply.github.com" ]
jnga@users.noreply.github.com
1f0aab49aa5a6590e8778e8b8366970e2e0a08f6
62babb33b9bede95aac217db04636956279bb2e2
/bit operation/1395C Boboniu and Bit Operations.py
90ae03a3fd60423b3df792021485ced2af7a8c6a
[]
no_license
tycyd/codeforces
0322e31daf18544944c769fd2a50c6d006015e34
e0773f069c6c5793f9d9a07b61878a589e375a5f
refs/heads/master
2023-08-12T05:00:39.467404
2021-09-30T16:39:21
2021-09-30T16:39:21
266,847,425
0
0
null
null
null
null
UTF-8
Python
false
false
589
py
from sys import stdin, stdout # 1 1 1 => 1 0 0, 0 1 1 # 1 1 0 0 => 1 0 0 # def boboniu_and_bit_operations(n, m, a_a, b_a): for k in range(513): cnt = 0 for a in a_a: for b in b_a: if ((a & b) | k) == k: cnt += 1 break if cnt == n: return k return -1 n, m = map(int, stdin.readline().split()) a_a = list(map(int, stdin.readline().split())) b_a = list(map(int, stdin.readline().split())) stdout.write(str(boboniu_and_bit_operations(n, m, a_a, b_a)) + '\n')
[ "tycyd@hotmail.com" ]
tycyd@hotmail.com
edf977c8ee2771f059d611fdf4b49337c5b6119e
a4174a9d51577d9b72b4e5dcf1be56bc9b0d242b
/retinanet/model/head/builder.py
b4153ffafb41099f951afdc540259b1454c0ab31
[ "Apache-2.0" ]
permissive
lchen-wyze/retinanet-tensorflow2.x
996396724c858fdc954880f3c20db7865d930a87
86404a2da6ec636d4b1aef768ac52f018c127798
refs/heads/master
2023-08-23T06:12:39.629288
2021-10-18T15:52:23
2021-10-18T15:52:23
418,040,957
0
0
Apache-2.0
2021-10-17T06:26:21
2021-10-17T06:26:21
null
UTF-8
Python
false
false
2,157
py
import numpy as np import tensorflow as tf from retinanet.model.head.detection_head import DetectionHead def build_detection_heads( params, min_level, max_level, conv_2d_op_params=None, normalization_op_params=None, activation_fn=None): if activation_fn is None: raise ValueError('`activation_fn` cannot be None') box_head = DetectionHead( num_convs=params.num_convs, filters=params.filters, output_filters=params.num_anchors * 4, min_level=min_level, max_level=max_level, prediction_bias_initializer='zeros', conv_2d_op_params=conv_2d_op_params, normalization_op_params=normalization_op_params, activation_fn=activation_fn, name='box-head') prior_prob_init = tf.constant_initializer(-np.log((1 - 0.01) / 0.01)) class_head = DetectionHead( num_convs=params.num_convs, filters=params.filters, output_filters=params.num_anchors*params.num_classes, min_level=min_level, max_level=max_level, prediction_bias_initializer=prior_prob_init, conv_2d_op_params=conv_2d_op_params, normalization_op_params=normalization_op_params, activation_fn=activation_fn, name='class-head') return box_head, class_head def build_auxillary_head( num_convs, filters, num_anchors, min_level, max_level, conv_2d_op_params=None, normalization_op_params=None, activation_fn=None): if activation_fn is None: raise ValueError('`activation_fn` cannot be None') prior_prob_init = tf.constant_initializer(-np.log((1 - 0.5) / 0.5)) auxillary_head = DetectionHead( num_convs=num_convs, filters=filters, output_filters=num_anchors, min_level=min_level, max_level=max_level, prediction_bias_initializer=prior_prob_init, conv_2d_op_params=conv_2d_op_params, normalization_op_params=normalization_op_params, activation_fn=activation_fn, name='auxillary-head') return auxillary_head
[ "sriharihumbarwadi97@gmail.com" ]
sriharihumbarwadi97@gmail.com
59f8819b81b135eec58e7e72e99829415c9a5917
2312db84a76baf32a918ed9d354dc324c842c3cb
/data-structure/array/max_ascending_pairs.py
168918681a9d68f94ef547f4f6c7a55cfcc5791a
[]
no_license
justdoit0823/notes
a6ec61c2429a76f7d9ac52d8535a97abfa96ee10
25da05f9fdd961949c56bcb11711934d53d50c9a
refs/heads/master
2023-06-22T23:17:17.548741
2022-10-19T15:15:15
2022-10-19T15:15:15
18,003,252
13
1
null
2023-06-14T22:29:39
2014-03-22T06:01:28
Jupyter Notebook
UTF-8
Python
false
false
1,103
py
"""Max ascending pairs.""" def max_ascending_pairs(a_list): m_low_idx = None m_high_idx = None c_low_idx = None c_high_idx = None for idx, val in enumerate(a_list): if idx == 0: c_low_idx = idx c_high_idx = idx else: if val >= a_list[idx - 1]: c_high_idx = idx else: if m_low_idx is None: m_low_idx = c_low_idx m_high_idx = c_high_idx else: if (c_high_idx - c_low_idx) > (m_high_idx - m_low_idx): m_low_idx = c_low_idx m_high_idx = c_high_idx c_low_idx = idx c_high_idx = idx if (c_high_idx - c_low_idx) > (m_high_idx - m_low_idx): m_low_idx = c_low_idx m_high_idx = c_high_idx return a_list[m_low_idx: m_high_idx + 1] def main(): test_list = [1, 2, 3, 2, 3, 4, 5, 6, 5, 7, 8, 9, 10, 11] assert max_ascending_pairs(test_list) == [5, 7, 8, 9, 10, 11] if __name__ == '__main__': main()
[ "justdoit920823@gmail.com" ]
justdoit920823@gmail.com
3a141b982aad58e4efe5283b889680634f9fbab5
9ca57420255bdc8df9f9d34989a28457ef818d66
/src/tests/test_era5_processor_service.py
306d90806bca697684582fcc3fff6676eb9d2d5c
[]
no_license
SatelliteApplicationsCatapult/csvs-nc-processor
6b8475bbbc2e415e718ff9b0acbf7c909fe7d0ba
c0ab5dfee88e78717d6589b77b1d64f2fe5b5f03
refs/heads/master
2023-01-13T08:57:09.991893
2020-10-23T14:49:51
2020-10-23T14:49:51
294,157,819
0
0
null
null
null
null
UTF-8
Python
false
false
1,006
py
import pathlib import unittest from services.era5_processor_service import merge_nc_files, MergeError, concatenate_nc_files extracted_daily_data_path = 'tests/resources/extracted_daily_data' extracted_monthly_data_path = 'tests/resources/extracted_monthly_data' class ERA5ProcessorService(unittest.TestCase): def test_merge_daily_data(self): extracted_daily_data = pathlib.Path(extracted_daily_data_path) daily_files = [str(x) for x in extracted_daily_data.glob('**/*.nc') if x.is_file()] merge_nc_files(daily_files, 'test_daily_data.nc') self.assertRaises(MergeError) def test_concatenate_monthly_data(self): extracted_monthly_data = pathlib.Path(extracted_monthly_data_path) monthly_files = [str(x) for x in extracted_monthly_data.glob('**/*.nc') if x.is_file()] concatenate_nc_files(monthly_files, 'test_monthly_data.nc') self.assertRaises(MergeError) if __name__ == '__main__': unittest.main()
[ "noreply@github.com" ]
noreply@github.com
a71ce795ba83eeadfd0ca322c34f7d911d52d157
92a4bd6df16659e72d91b279f915d81b84d56c39
/GBN_window.py
bccc5c1cd549160c1a7459a268474e5ffdc5bc91
[]
no_license
banana16314/rdt
1d1d714dc34554065f37de59da5abf24fac4963c
1f1ca293c0c2462861997f0728a2bc9644f9a315
refs/heads/master
2020-12-30T15:29:41.671132
2017-05-19T04:39:12
2017-05-19T04:39:12
91,145,369
0
0
null
null
null
null
UTF-8
Python
false
false
113
py
class GBN_window : def __init__(self) : self.packet_list = [] self.ack_list = [] self.resend_list = []
[ "18245032253@163.com" ]
18245032253@163.com
376b7b79d23557f1e0c36c97dd6bb44441e25287
902fb2fdc99eb16df86344bee0ac5ac02d1686a9
/lib/my_requests.py
849dad0edf3f5591da57b5cf3bd0b20934093ae8
[]
no_license
Natasha093/LearnQA_PythonAPI2
ae665705af07f9752871004a9da247aaa1ecfac8
343ead661b0b54f6f2160381d77074b6824528c8
refs/heads/master
2023-07-09T05:28:30.287901
2021-08-17T11:24:49
2021-08-17T11:24:49
393,113,137
0
0
null
null
null
null
UTF-8
Python
false
false
2,043
py
import requests from lib.logger import Logger import allure from environment import ENV_OBJECT class MyRequests: @staticmethod def post(url: str, data: dict = None, headers: dict = None, cookies: dict = None): with allure.step(f"POST request to URL '{url}'"): return MyRequests._send(url, data, headers, cookies, 'POST') @staticmethod def get(url: str, data: dict = None, headers: dict = None, cookies: dict = None): with allure.step(f"GET request to URL '{url}'"): return MyRequests._send(url, data, headers, cookies, 'GET') @staticmethod def put(url: str, data: dict = None, headers: dict = None, cookies: dict = None): with allure.step(f"PUT request to URL '{url}'"): return MyRequests._send(url, data, headers, cookies, 'PUT') @staticmethod def delete(url: str, data: dict = None, headers: dict = None, cookies: dict = None): with allure.step(f"DELETE request to URL '{url}'"): return MyRequests._send(url, data, headers, cookies, 'DELETE') @staticmethod def _send(url: str, data: dict, headers: dict, cookies: dict, method: str): url = f"{ENV_OBJECT.get_base_url()}{url}" if headers is None: headers = {} if cookies is None: cookies = {} Logger.add_request(url, data, headers, cookies, method) if method == 'GET': response = requests.get(url, params=data, headers=headers, cookies=cookies) elif method == 'POST': response = requests.post(url, data=data, headers=headers, cookies=cookies) elif method == 'PUT': response = requests.put(url, data=data, headers=headers, cookies=cookies) elif method == 'DELETE': response = requests.delete(url, data=data, headers=headers, cookies=cookies) else: raise Exception(f"Bad HTTP method '{method}' was received") Logger.add_response(response) return response
[ "fanloko063@mail.ru" ]
fanloko063@mail.ru
f668e324466fcbe5eacf8b3b1dfb24bd51d73cfd
5ef9d5114cd6c16c22f289f65f86de16bc15e46b
/src/smartcity/module/modelling/arima/x1ClassifyTrafficArimaData.py
fbf39b98f135f78e0dd10f28cb0309f40dceefeb
[]
no_license
decmhugh/smartertraffic
d30b79d4d8d602741c2e8aa368cce7886b11558f
b672c4cd1ce26e2c4b58f3799da978ab890b5ece
refs/heads/master
2021-01-02T22:30:51.091589
2014-10-21T09:04:41
2014-10-21T09:04:41
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,369
py
''' Created on 3 Dec 2013 @author: declan ''' from pymongo import Connection as mongoConn import os,gc,sys import numpy as np from datetime import datetime, date from pandas.core.series import TimeSeries from sklearn import cross_validation from pandas import DataFrame as df from sklearn import linear_model from bson import json_util import pandas as pd from sklearn import svm from scipy import stats import statsmodels.formula.api as sm from statsmodels.graphics.api import qqplot from statsmodels.graphics.tsaplots import plot_acf from statsmodels.graphics.tsaplots import plot_pacf import matplotlib.pyplot as plt from statsmodels.graphics.tsaplots import plot_acf, plot_pacf connection = mongoConn('mongodb://localhost:27017/') db = connection.traffic classifier_list = [ linear_model.LinearRegression(), linear_model.ElasticNet(), linear_model.PassiveAggressiveRegressor() ] def process_lock(args): data = df().from_csv("pickle/" + args.replace("/","_") + "/" + "data.csv") data = data.resample('B').dropna() #data = df({"STT":[1,2,3,4,5,6,7,5,3,2,2,2],"S":[1,2,3,4,5,6,7,5,3,2,2,2]}) #plt.plot(data["STT"].values) d = df({"STT":data["STT"].values, "STT1":data.shift(1)["STT"], "STT2":data.shift(2)["STT"], "STT3":data.shift(3)["STT"], "STT4":data.shift(4)["STT"], "STT5":data.shift(5)["STT"], "STT6":data.shift(6)["STT"]}) d = d.dropna().astype(int) x = d.copy() x.__delitem__("STT") #print(x) #d.plot(figsize=(12,8)); #plt.plot(RangeData.range(d)) fig = plt.figure(figsize=(12,8)) ax1 = fig.add_subplot(211) fig = plot_acf(d["STT"].values.squeeze(), lags=40, ax=ax1) #fig = plot_pacf(d, lags=40, ax=ax1) #d.plot() plt.show() #ax2 = fig.add_subplot(212) #fig = plot_pacf(d.index.values, lags=40, ax=ax2) #print("Get training data") #X_train, X_test, y_train, y_test = cross_validation.train_test_split( # x, d["STT"].values, test_size=0.4, random_state=0) #print(X_train.shape, y_train.shape) #print(X_test.shape, y_test.shape) print("-----------SCORE-----------------------------") print(args,">>>") result = sm.ols(formula="STT ~ STT1 + STT2 + STT3 + STT4 + STT5 + STT6", data=d).fit() print(result.summary()) #fig = plt.figure(figsize=(12,8)) #ax1 = fig.add_subplot(211) #linear_model. #fig = sm.graphics.tsa.plot_acf(d["STT"].values, lags=40, ax=ax1) #ax2 = fig.add_subplot(212) #fig = sm.graphics.tsa.plot_pacf(d["STT"].values, lags=40, ax=ax2) #print("Get Classifier") #clf = svm.SVC(kernel='linear', C=1).fit(X_train, y_train) #print("Get Score") #print(clf.score(X_test, y_test)) #predicted = clf.predict(X_test) #print(predicted) #scores = cross_validation.cross_val_score(clf, X_test, y_test, cv=6) #print(scores) #print(scores.std()) #print("Accuracy: %0.2f (+/- %0.2f)" % (scores.mean(), scores.std() * 2)) #print("Explained Variance Score: %0.2f" % explained_variance_score(y_test, predicted)) #print("Mean Absolute Squared: %0.2f" % mean_absolute_error(y_test, predicted)) #print("Mean Squared Squared: %0.2f" % mean_squared_error(y_test, predicted)) def process(args): pickle_dir = "pickle/" + args.replace("/","_") print(pickle_dir) result = None if not os.path.isdir(pickle_dir): os.makedirs(pickle_dir) a = datetime.now().strftime('%Y%m%d%H%M%S') try: process_lock(args); #print(result) #pickle.dump(result, open(pickle_dir + "/result.res", 'wb')) except ValueError as error: print(error) b = datetime.now().strftime('%Y%m%d%H%M%S') with open("pickle/" + args.replace("/","_") + "/timestamp.txt", "w") as target: target.write("start:" + a) target.write("end:" + b) target.close() return result; def run(): cursor = db.junctions.find() json_str =json_util.dumps(cursor) junctions =json_util.loads(json_str) junctions = sorted(junctions, key=lambda k: k['route']) for junction in junctions: process(junction["_id"]) if __name__ == "__main__": for p in ["17/6/1","13/2/1","30/7/1","10/7/2","16/2/2","30/4/2"]: process_lock(p)
[ "declan.mchugh@gmail.com" ]
declan.mchugh@gmail.com