blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
288
content_id
stringlengths
40
40
detected_licenses
listlengths
0
112
license_type
stringclasses
2 values
repo_name
stringlengths
5
115
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
684 values
visit_date
timestamp[us]date
2015-08-06 10:31:46
2023-09-06 10:44:38
revision_date
timestamp[us]date
1970-01-01 02:38:32
2037-05-03 13:00:00
committer_date
timestamp[us]date
1970-01-01 02:38:32
2023-09-06 01:08:06
github_id
int64
4.92k
681M
star_events_count
int64
0
209k
fork_events_count
int64
0
110k
gha_license_id
stringclasses
22 values
gha_event_created_at
timestamp[us]date
2012-06-04 01:52:49
2023-09-14 21:59:50
gha_created_at
timestamp[us]date
2008-05-22 07:58:19
2023-08-21 12:35:19
gha_language
stringclasses
147 values
src_encoding
stringclasses
25 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
128
12.7k
extension
stringclasses
142 values
content
stringlengths
128
8.19k
authors
listlengths
1
1
author_id
stringlengths
1
132
15f8cd0d04070fe3d412a6539a5235a6a8ef6c98
da19363deecd93a73246aaea877ee6607daa6897
/xlsxwriter/test/worksheet/test_write_filter.py
bfe6d60fc98f85a51ac98245e67242f2736abce1
[]
no_license
UNPSJB/FarmaciaCrisol
119d2d22417c503d906409a47b9d5abfca1fc119
b2b1223c067a8f8f19019237cbf0e36a27a118a6
refs/heads/master
2021-01-15T22:29:11.943996
2016-02-05T14:30:28
2016-02-05T14:30:28
22,967,417
0
0
null
null
null
null
UTF-8
Python
false
false
775
py
############################################################################### # # Tests for XlsxWriter. # # Copyright (c), 2013-2015, John McNamara, jmcnamara@cpan.org # import unittest from ...compatibility import StringIO from ...worksheet import Worksheet class TestWriteFilter(unittest.TestCase): """ Test the Worksheet _write_filter() method. """ def setUp(self): self.fh = StringIO() self.worksheet = Worksheet() self.worksheet._set_filehandle(self.fh) def test_write_filter(self): """Test the _write_filter() method""" self.worksheet._write_filter('East') exp = """<filter val="East"/>""" got = self.fh.getvalue() self.assertEqual(got, exp)
[ "lealuque.tw@gmail.com" ]
lealuque.tw@gmail.com
9db96c6dbb62656d25dbde9c5f8ba667c8585fa8
3817a595bdf246996c1f3846860cea9a9154f1cc
/AltviaDogs/apps.py
6e7c68cfaf6935515b6abbaa873e7c0dd3d098b0
[]
no_license
shedwyn/AltviaDogs
99e83f217d891c5060d0b9a8f0997ea4ea2c5dc2
150078897d38c84ac3c492635ea777476e84b0d9
refs/heads/master
2021-01-24T00:32:32.176410
2018-08-24T20:01:47
2018-08-24T20:01:47
122,768,745
0
0
null
null
null
null
UTF-8
Python
false
false
319
py
""" Designed and built by Erin 'Ernie' L. Fough. June 2018 (start) contact: shedwyn@gmail.com want to know when this was last updated? See README.md. """ from django.apps import AppConfig class AltviaDogsConfig(AppConfig): """Establish Link to Dog Manager (AltviaDogs) Application.""" name = 'AltviaDogs'
[ "shedwyn@gmail.com" ]
shedwyn@gmail.com
23bcdebc6312f35fb67216dff2380ccbe74890ab
68215f224dd76f8bf299c66c80cb4a0de1a05a20
/PythonTutorial/Advance/MetaClass/meta_class_example2.py
2e013f02ffac5c75fa0acaf6a410b099a0bd7425
[]
no_license
jhappye/Python-Study
3935e1d9ffead3afb1b39823a4114006094a6542
40a421980e9f56a97f6f7b9eb00bbf983ee12c5b
refs/heads/master
2023-08-30T17:36:26.313896
2021-11-16T09:53:39
2021-11-16T09:53:39
null
0
0
null
null
null
null
UTF-8
Python
false
false
226
py
# 常规方法创建 class Foo: pass # 常规方法创建 class Bar(Foo): attr = 100 x = Foo() print(x) # 动态创建,和上面等价 Bar = type('Bar', (Foo,), dict(attr=100)) x = Bar() print(x) print(type(x))
[ "516495459@qq.com" ]
516495459@qq.com
76a14a2500ada13704b7d69ed40d39c206497011
680db028bdfd570688e66009dd41a424b2395a6e
/Camelcase.py
51a76dd0a42b6afcae6e169004a153a3dac0257b
[]
no_license
vitcmaestro/player
46fcf3138a7ba98a3aadae60aebdd68efdd16c55
8b9b132ec5f3dfe22521daf6122431451db8d880
refs/heads/master
2020-04-15T16:43:53.944347
2019-01-20T05:00:31
2019-01-20T05:00:31
164,848,284
1
1
null
null
null
null
UTF-8
Python
false
false
205
py
str1 = input("") str2 ="" for i in range(len(str1)): if(i==0): str2 +=str1[i].upper() elif(str1[i-1].isspace()): str2 +=str1[i].upper() else: str2 +=str1[i] print(str2)
[ "noreply@github.com" ]
vitcmaestro.noreply@github.com
149488d35576cf1a5937c6f95ae55ab3c88828d9
837d683ccf6aa08b0ad6d95f6af47fcc12b72da3
/pedersen_vss.py
586d0dff4ce73808e4596e669d77272e83f15f10
[]
no_license
froyobin/secretsharing
deef2cdb9be72ff51be339bfaae4822e351b8cf1
737f2aa074e0cb6f082916bd11afc658732777e1
refs/heads/master
2021-01-11T21:32:12.720629
2017-01-20T00:46:31
2017-01-20T00:46:31
78,798,569
2
0
null
null
null
null
UTF-8
Python
false
false
4,479
py
import argparse import shamir_secret_sharing from Crypto.Util import number #To make it simple, we give the f(X)=5+3x+8x^2 as params. It is easy to do # the test and write the code. p = int( "FFFFFFFFFFFFFFFFC90FDAA22168C234C4C6628B80DC1CD129024E088A67CC74020BBEA63B139B22514A08798E3404DDEF9519B3CD3A431B302B0A6DF25F14374FE1356D6D51C245E485B576625E7EC6F44C42E9A637ED6B0BFF5CB6F406B7EDEE386BFB5A899FA5AE9F24117C4B1FE649286651ECE45B3DC2007CB8A163BF0598DA48361C55D39A69163FA8FD24CF5F83655D23DCA3AD961C62F356208552BB9ED529077096966D670C354E4ABC9804F1746C08CA18217C32905E462E36CE3BE39E772C180E86039B2783A2EC07A28FB5C55DF06F4C52C9DE2BCBF6955817183995497CEA956AE515D2261898FA051015728E5A8AACAA68FFFFFFFFFFFFFFFF", 16) g = 2 h = int( "3d941d6d9cd4c77719840d6b391a63ca6ded5b5cf6aafeefb9ea530f523039e9c372736a79b7eb022e50029f7f2cb4fb16fd1def75657288eca90d2c880f306be76fe0341b3c8961ae6e61aabbb60e416069d97eeada2f1408f2017449dddcd5ac927f164b1a379727941bd7f2170d02ef12ef3ec801fae585ac7b9d4079f50feced64687128208d46e3e10c5d78eb05832f5322c07a3b4e14c6f595206fde99115e8eea19b5fb13dd434332ec3eccb41a4baa54a14183c3416313678697db8507abdcfc6a97c86099fa5172316d784c6997fc2e74e8e59c7c1bc90426164682f5bfbf6373b13ea90d7e13fbffd65e10c4ad96c38ccbf8e8def28d76746729dc", 16) def create_verifies(params1, params2, p, g, h): verifies = [] for i in range(0, len(params1)): verifies.append((pow(g, params1[i], p) * pow(h, params2[i], p)) % p) return verifies def verify_each(i, verifies, len_pamras): left_value = 1 for j in range(0, len_pamras): upper = pow(i, j, p) left_value *= pow(verifies[j], upper, p) def calculate_left(verfies, i, t, p): powerall = [1] for each_t in range(1, t): powerall.append(pow(i, each_t)) left_val = 1 for j in range(0, len(verfies)): c = pow(verfies[j], powerall[j], p) left_val *= c left_val %= p return left_val % p def verifies_shares(secrets1, secrets2, verifies, params,p,g): for i in range(0, len(secrets1)): left_value = calculate_left(verifies, i+1, len(params), p) right_value = (pow(g, (secrets1[i])%p, p)*pow(h, (secrets2[i])%p, p))%p if left_value == right_value: print "checking %d Successfully!!" % i else: print "secret %d has been modified!!" % i def test_prime(): n_length = 2048 q= (p-1)%2 print q print number.isPrime((p-1)/2) primeNum1 = number.getPrime(n_length) # primeNum2 = number.getStrongPrime(n_length, primeNum1) i=2 while True: pl = p-1 print i if pl%i == 0: if number.isPrime(pl / i): print "############" print i print "############" break i += 1 if i==10000: break print "found!!" print "############################" print primeNum1 print "############################" def main(args): s = shamir_secret_sharing.StringtoInt(args.secret) parties = int(args.party) min_party = int(args.min_party) params1 = shamir_secret_sharing.create_params(s, min_party, p) params2 = shamir_secret_sharing.create_params(s, min_party, p) secrets1 = shamir_secret_sharing.create_secret(parties, params1, p) secrets2 = shamir_secret_sharing.create_secret(parties, params2, p) secret = shamir_secret_sharing.construct_secret(secrets1, min_party, p) verifies = create_verifies(params1, params2, p, g, h) verifies_shares(secrets1, secrets2, verifies, params2, p, g) print "The secret you give is " + str(params1[0]) + "\n" print "The rejoin code is " + str(secret) if params1[0] == secret: print "rejoin Successfully!!" else: print "we cannot rejoin the secret" #We change secret2 3's secret and see whether we can check it out secrets2[3] -= 1 verifies_shares(secrets1, secrets2, verifies, params2, p, g) if __name__ == "__main__": parser = argparse.ArgumentParser(description='Create secret shares') parser.add_argument('secret', metavar='\"secret\"', type=str, help='the secret to share') parser.add_argument('party', metavar='\"secret\"', type=str, help='the secret to share') parser.add_argument('min_party', metavar='\"secret\"', type=str, help='the secret to share') args = parser.parse_args() main(args)
[ "froyo.bin@gmail.com" ]
froyo.bin@gmail.com
32e1da53f66f709ec826c5e6e98b69f964055011
644bcdabf35261e07c2abed75986d70f736cb414
/python-project/Defis/Euler_44_Test.py
059d736f3c4546ec361ff33f1fc9b32183c5bd48
[]
no_license
matcianfa/playground-X1rXTswJ
f967ab2c2cf3905becafb6d77e89a31414d014de
67859b496e407200afb2b1d2b32bba5ed0fcc3f0
refs/heads/master
2023-04-03T11:56:15.878757
2023-03-24T15:52:37
2023-03-24T15:52:37
122,226,979
5
20
null
null
null
null
UTF-8
Python
false
false
1,110
py
#Ne pas oublier de changer le module à importer module="Defis/Euler_44" import sys import io #On récupère les données de l'utilisateur sauvegarde_stdout=sys.stdout sys.stdout=io.StringIO() from Euler_44 import * count1 = sys.stdout.getvalue()[:-1] sys.stdout=sauvegarde_stdout from ma_bao import * #La réponse reponse=5482660 #message d'aide si besoin help="N'oublie pas d'utiliser print pour afficher le resultat" def send_msg(channel, msg): print("TECHIO> message --channel \"{}\" \"{}\"".format(channel, msg)) def success(): send_msg("Tests validés","Bravo !") afficher_correction(module) print("TECHIO> success true") def fail(): print("TECHIO> success false") def test(): try: assert str(count1) == str(reponse), "Le résultat obtenu est {} mais ce n'est pas le bon.".format(str(count1)) send_msg("Tests validés","Le résultat cherché est bien {}".format(str(count1))) success() except AssertionError as e: fail() send_msg("Oops! ", e) if help: send_msg("Aide 💡", help) if __name__ == "__main__": test()
[ "noreply@github.com" ]
matcianfa.noreply@github.com
07e8af6fb34efbf28fd13da8b7998358fa68674c
f9c2e53c48839fe5583f6da48e800eba793f83c8
/custom_components/hacs/repositories/integration.py
8470873279f13b52e25cd0fb77a631d30b168ab2
[]
no_license
Bobo2012/homeassistant-config
d6a5ebdd8de978d14a5d060ad77975280efdf95e
ee2567182e3ef250ee2a2e1a403fbc6f27057861
refs/heads/master
2020-08-21T20:04:53.531610
2019-10-18T22:59:54
2019-10-18T22:59:54
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,501
py
"""Class for integrations in HACS.""" import json from aiogithubapi import AIOGitHubException from homeassistant.loader import async_get_custom_components from .repository import HacsRepository, register_repository_class @register_repository_class class HacsIntegration(HacsRepository): """Integrations in HACS.""" category = "integration" def __init__(self, full_name): """Initialize.""" super().__init__() self.information.full_name = full_name self.information.category = self.category self.manifest = None self.domain = None self.content.path.remote = "custom_components" self.content.path.local = self.localpath @property def localpath(self): """Return localpath.""" return f"{self.system.config_path}/custom_components/{self.domain}" @property def config_flow(self): """Return bool if integration has config_flow.""" if self.manifest is not None: if self.information.full_name == "custom-components/hacs": return False return self.manifest.get("config_flow", False) return False async def validate_repository(self): """Validate.""" await self.common_validate() # Attach repository if self.repository_object is None: self.repository_object = await self.github.get_repo( self.information.full_name ) # Custom step 1: Validate content. if self.repository_manifest: if self.repository_manifest.content_in_root: self.content.path.remote = "" if self.content.path.remote == "custom_components": ccdir = await self.repository_object.get_contents( self.content.path.remote, self.ref ) if not isinstance(ccdir, list): self.validate.errors.append("Repostitory structure not compliant") for item in ccdir or []: if item.type == "dir": self.content.path.remote = item.path break if self.repository_manifest.zip_release: self.content.objects = self.releases.last_release_object.assets else: self.content.objects = await self.repository_object.get_contents( self.content.path.remote, self.ref ) self.content.files = [] for filename in self.content.objects or []: self.content.files.append(filename.name) if not await self.get_manifest(): self.validate.errors.append("Missing manifest file.") # Handle potential errors if self.validate.errors: for error in self.validate.errors: if not self.system.status.startup: self.logger.error(error) return self.validate.success async def registration(self): """Registration.""" if not await self.validate_repository(): return False # Run common registration steps. await self.common_registration() # Get the content of the manifest file. await self.get_manifest() # Set local path self.content.path.local = self.localpath async def update_repository(self): """Update.""" await self.common_update() # Get integration objects. if self.repository_manifest: if self.repository_manifest.content_in_root: self.content.path.remote = "" if self.content.path.remote == "custom_components": ccdir = await self.repository_object.get_contents( self.content.path.remote, self.ref ) if not isinstance(ccdir, list): self.validate.errors.append("Repostitory structure not compliant") self.content.path.remote = ccdir[0].path try: self.content.objects = await self.repository_object.get_contents( self.content.path.remote, self.ref ) except AIOGitHubException: return self.content.files = [] for filename in self.content.objects or []: self.content.files.append(filename.name) await self.get_manifest() # Set local path self.content.path.local = self.localpath async def reload_custom_components(self): """Reload custom_components (and config flows)in HA.""" self.logger.info("Reloading custom_component cache") del self.hass.data["custom_components"] await async_get_custom_components(self.hass) async def get_manifest(self): """Get info from the manifest file.""" manifest_path = f"{self.content.path.remote}/manifest.json" try: manifest = await self.repository_object.get_contents( manifest_path, self.ref ) manifest = json.loads(manifest.content) except Exception: # pylint: disable=broad-except return False if manifest: self.manifest = manifest self.information.authors = manifest["codeowners"] self.domain = manifest["domain"] self.information.name = manifest["name"] self.information.homeassistant_version = manifest.get("homeassistant") # Set local path self.content.path.local = self.localpath return True return False
[ "arsaboo@gmx.com" ]
arsaboo@gmx.com
d95810328bbed890c7daf776a9675beedc3d8cd2
473568bf080e3637ee118b374f77e9f561286c6c
/SudoPlacementCourse/LeadersInAnArray.py
b2b997ace46a797ad0dd241ff7cd5667a35adba3
[]
no_license
VineetPrasadVerma/GeeksForGeeks
c2f7fc94b0a07ba146025ca8a786581dbf7154c8
fdb4e4a7e742c4d67015977e3fbd5d35b213534f
refs/heads/master
2020-06-02T11:23:11.421399
2020-01-07T16:51:18
2020-01-07T16:51:18
191,138,925
0
0
null
null
null
null
UTF-8
Python
false
false
1,241
py
# n = int(input()) # # for k in range(n): # final_ans_list = [] # size_of_array = int(input()) # list_of_elements = input() # # int_list_of_elements = [int(i) for i in list_of_elements.split()] # # for i in range(len(int_list_of_elements)): # if i == len(int_list_of_elements) - 1: # final_ans_list.append(int_list_of_elements[i]) # break # # max_element = int_list_of_elements[i] # temp_max_element = max(int_list_of_elements[i+1:]) # # if max_element >= temp_max_element: # final_ans_list.append(int_list_of_elements[i]) # # string_list_of_elements = [str(i) for i in final_ans_list] # print(" ".join(string_list_of_elements)) n = int(input()) for _ in range(n): final_ans_list = [] size_of_array = int(input()) list_of_elements = input() int_list_of_elements = [int(i) for i in list_of_elements.split()] maximum = -1 for i in range(len(int_list_of_elements)-1, -1, -1): if int_list_of_elements[i] >= maximum: maximum = int_list_of_elements[i] final_ans_list.append(maximum) for i in range(len(final_ans_list)-1, -1, -1): print(final_ans_list[i], end=" ") print()
[ "vineetpd1996@gmail.com" ]
vineetpd1996@gmail.com
8294fce530b848259f949592831c9de1c760dbad
14a913fce4b538b22f28409645cd6abe3455808f
/bigquery/cloud-client/natality_tutorial.py
5bfa8f1d27a9736075f271b77657e9342df6d688
[ "LicenseRef-scancode-unknown-license-reference", "Apache-2.0" ]
permissive
iamLoi/Python-Random-Number-Generator
8da7dbd37cb13a01232c8ed49b9df35a99c63d73
7579e8b15130802aaf519979e475c6c75c403eda
refs/heads/master
2022-08-29T19:05:32.649931
2019-09-14T14:48:58
2019-09-14T14:48:58
208,454,877
2
1
Apache-2.0
2022-08-05T21:57:49
2019-09-14T14:51:05
Python
UTF-8
Python
false
false
2,926
py
#!/usr/bin/env python # Copyright 2018 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. def run_natality_tutorial(): # [START bigquery_query_natality_tutorial] """Create a Google BigQuery linear regression input table. In the code below, the following actions are taken: * A new dataset is created "natality_regression." * A query is run against the public dataset, bigquery-public-data.samples.natality, selecting only the data of interest to the regression, the output of which is stored in a new "regression_input" table. * The output table is moved over the wire to the user's default project via the built-in BigQuery Connector for Spark that bridges BigQuery and Cloud Dataproc. """ from google.cloud import bigquery # Create a new Google BigQuery client using Google Cloud Platform project # defaults. client = bigquery.Client() # Prepare a reference to a new dataset for storing the query results. dataset_ref = client.dataset('natality_regression') dataset = bigquery.Dataset(dataset_ref) # Create the new BigQuery dataset. dataset = client.create_dataset(dataset) # In the new BigQuery dataset, create a reference to a new table for # storing the query results. table_ref = dataset.table('regression_input') # Configure the query job. job_config = bigquery.QueryJobConfig() # Set the destination table to the table reference created above. job_config.destination = table_ref # Set up a query in Standard SQL, which is the default for the BigQuery # Python client library. # The query selects the fields of interest. query = """ SELECT weight_pounds, mother_age, father_age, gestation_weeks, weight_gain_pounds, apgar_5min FROM `bigquery-public-data.samples.natality` WHERE weight_pounds IS NOT NULL AND mother_age IS NOT NULL AND father_age IS NOT NULL AND gestation_weeks IS NOT NULL AND weight_gain_pounds IS NOT NULL AND apgar_5min IS NOT NULL """ # Run the query. query_job = client.query(query, job_config=job_config) query_job.result() # Waits for the query to finish # [END bigquery_query_natality_tutorial] if __name__ == '__main__': run_natality_tutorial()
[ "noreply@github.com" ]
iamLoi.noreply@github.com
0618850a5db7746378ebd7c6d18fc84b9305a5e3
05324b134108e0e0fde392e0ae0fc22bfa1fb75f
/df_user/islogin.py
d2149c4633a4871adc7f90b232f8ccc8e25afcec
[]
no_license
2339379789/zhang_ttshop
2c8d546b9ed3710fd1f48d6075ea01955247d34f
44f9eb998182f4aa027d5d313b4957410b54a39d
refs/heads/master
2020-03-06T15:40:54.261769
2018-03-27T09:23:52
2018-03-27T09:23:52
126,960,140
2
1
null
null
null
null
UTF-8
Python
false
false
387
py
from django.http import HttpResponseRedirect def islogin(func): def login_fun(request, *args, **kwargs): if request.session.get('user_id'): return func(request, *args, **kwargs) else: red = HttpResponseRedirect('/user/login') red.set_cookie('url', request.get_full_path) return red return login_fun
[ "2339379789@qq.com.com" ]
2339379789@qq.com.com
fcdd844f805ea4a7bd79824935397e4ae355b4f3
62e58c051128baef9452e7e0eb0b5a83367add26
/x12/4052/354004052.py
c318eb1028c5e6c2d9a7772e6ec4c9f7cf7069da
[]
no_license
dougvanhorn/bots-grammars
2eb6c0a6b5231c14a6faf194b932aa614809076c
09db18d9d9bd9d92cefbf00f1c0de1c590fe3d0d
refs/heads/master
2021-05-16T12:55:58.022904
2019-05-17T15:22:23
2019-05-17T15:22:23
105,274,633
0
0
null
2017-09-29T13:21:21
2017-09-29T13:21:21
null
UTF-8
Python
false
false
453
py
from bots.botsconfig import * from records004052 import recorddefs syntax = { 'version' : '00403', #version of ISA to send 'functionalgroup' : 'AY', } structure = [ {ID: 'ST', MIN: 1, MAX: 1, LEVEL: [ {ID: 'M10', MIN: 1, MAX: 1}, {ID: 'P4', MIN: 1, MAX: 20, LEVEL: [ {ID: 'X01', MIN: 1, MAX: 1}, {ID: 'X02', MIN: 0, MAX: 9999}, ]}, {ID: 'SE', MIN: 1, MAX: 1}, ]} ]
[ "jason.capriotti@gmail.com" ]
jason.capriotti@gmail.com
c3b06571705e3d42ea421dbcc6cb94538267f247
078533ce919371451564646e3c311c8dd6fca7ea
/app/settings/prod.py
1425b25dbaa062e89c3c39a5947da1224923cd03
[]
no_license
cr8ivecodesmith/djdraft
cf5415a967dc7fc6f4f8d191def9c2b687c0d744
364ded3ea43acc874de367cd679c4bddfb64d837
refs/heads/master
2020-04-05T23:07:20.690873
2016-10-06T05:11:29
2016-10-06T05:11:29
23,010,461
2
1
null
2016-10-03T13:42:04
2014-08-16T04:17:23
Python
UTF-8
Python
false
false
1,558
py
from .base import * DEBUG = False ALLOWED_HOSTS = [ '{{ project_name }}.dev', ] ###### APP CONFIGURATION # See: https://docs.djangoproject.com/en/dev/ref/settings/#installed-apps INSTALLED_APPS += [ ] ###### DATABASE CONFIGURATION # https://docs.djangoproject.com/en/dev/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql_psycopg2', 'NAME': 'caffeinedb', 'USER': 'caffeine', 'PASSWORD': 'changeme', 'HOST': 'postgres', 'PORT': 5432, } } ########## EMAIL CONFIGURATION # See: https://docs.djangoproject.com/en/dev/topics/email/#smtp-backend EMAIL_BACKEND = 'django_smtp_ssl.SSLEmailBackend' # See: https://docs.djangoproject.com/en/dev/ref/settings/#email-host-user EMAIL_HOST_USER = get_key('EMAIL_HOST_USER') # See: https://docs.djangoproject.com/en/dev/ref/settings/#email-host-password EMAIL_HOST_PASSWORD = get_key('EMAIL_HOST_PASSWORD') # See: https://docs.djangoproject.com/en/dev/ref/settings/#email-host EMAIL_HOST = 'smtp.{{ project_name }}.dev' # See: https://docs.djangoproject.com/en/dev/ref/settings/#email-port EMAIL_PORT = 465 SERVER_EMAIL = 'errors@{{ project_name }}.dev' DEFAULT_FROM_EMAIL = 'noreply@{{ project_name }}.dev' ###### MANAGER CONFIGURATION # See: https://docs.djangoproject.com/en/dev/ref/settings/#admins ADMINS += [ ('caffeine prod', 'errors@{{ project_name }}.dev'), ] ###### CELERY CONFIGURATION # CELERY_ROUTES = { # 'app.tasks.example_queue': { # 'queue': 'express_queue' # }, # }
[ "matt@lebrun.org" ]
matt@lebrun.org
0af0e6aeb4bf93bb5c2c00acba6daf1feb4788c2
0566cf76b456518875edecece15e763a36a4795f
/scrapers/tv_showsonline_com.py
6208f8a041fec633df9bb9e2989a3db26e001ce2
[]
no_license
theclonedude/Scraping_BeautifulSoup_phantomjs
684b1f7a993e0d2555daa7a5455cf19bd29b0b1b
faf653feae46c21a72d13b2123cdebdb2f7c05d8
refs/heads/master
2023-03-16T19:36:14.867361
2018-06-14T14:21:02
2018-06-14T14:21:02
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,148
py
from sandcrawler.scraper import ScraperBase from sandcrawler.scraper import SimpleScraperBase class TvShowsOnline(SimpleScraperBase): BASE_URL = 'http://www.7stream.pro/' OTHER_URLS = ['http://tv-showsonline.com', ] def setup(self): self.register_scraper_type(ScraperBase.SCRAPER_TYPE_OSP) self.search_term_language = 'eng' raise NotImplementedError('the website returns the bad gateway error') self.register_media(ScraperBase.MEDIA_TYPE_FILM) self.register_media(ScraperBase.MEDIA_TYPE_TV) for url in [self.BASE_URL, ] + self.OTHER_URLS: self.register_url(ScraperBase.URL_TYPE_SEARCH, url) self.register_url(ScraperBase.URL_TYPE_LISTING, url) def _fetch_search_url(self, search_term, media_type): return self.BASE_URL + '/?s='+ '%s' % self.util.quote(search_term) def _fetch_no_results_text(self): return 'No Articles Found' def _fetch_next_button(self, soup): link = soup.select_one('a[class="next page-numbers"]') self.log.debug(link) self.log.debug('------------------------') return link['href'] if link else None def _parse_search_result_page(self, soup): results = soup.select('div.item_content h4 a') if not results or len(results) == 0: return self.submit_search_no_results() for result in results: self.submit_search_result( link_url = result['href'], link_title = result.text ) def _parse_parse_page(self, soup): title = soup.select_one('h1[class="entry_title entry-title"]').text.strip() season, episode = self.util.extract_season_episode(title) titles = soup.select('span.ser-name') index = 0 for link in soup.select('a.wo-btn'): self.submit_parse_result( index_page_title = self.util.get_page_title(soup), link_url=link['href'], link_title=titles[index].text, series_season=season, series_episode=episode ) index += 1
[ "stryokka@gmail.com" ]
stryokka@gmail.com
585c1f38e68fa896e9f1b91523d7cde15f833d05
544fe02a27cc4d987724b1bf45c2ba2994676521
/Q6.3_brain_teasers.py
c42fb93923fdbb832381c85a4fbe252f28b128c3
[ "Unlicense" ]
permissive
latika18/learning
1e7a6dbdea399b845970317dc62089911a13df1c
a57c9aacc0157bf7c318f46c1e7c4971d1d55aea
refs/heads/master
2021-06-16T19:20:28.146547
2019-09-03T06:43:28
2019-09-03T06:43:28
115,537,386
0
0
null
null
null
null
UTF-8
Python
false
false
524
py
#You have a five quart jug and a three quart jug, and an unlimited supply of water (but no measuring cups). How would you come up with exactly four quarts of water? #NOTE: The jugs are oddly shaped, such that filling up exactly ‘half’ of the jug would be impossible. _ ________________________________________________________________ 5 Quart Contents 3 Quart Contents Note 5 0 Filled 5 quart jug 2 3 Filled 3Q with 5Q’s contents 0 2 Dumped 3Q 5 2 Filled 5Q 4 3 Fill remainder of 3Q with 5Q 4 Done! We have four quarts
[ "noreply@github.com" ]
latika18.noreply@github.com
e2e7d5b9b92f7a0b1b6801b43db9c95bca9229f0
1d159ff6d4d72b1a2399916ec1e28ef885b59323
/solutions/module_3/01_guess_my_number.py
4b9c9b119b0c1c64362318c3850d074b0c21f484
[]
no_license
Zerl1990/python_essentials
4e329b6e36b36ff340d505b26608d2b244ad2d09
ce257c25072debed9717960591e39c586edd7110
refs/heads/master
2023-02-28T17:42:17.465769
2021-02-02T23:39:41
2021-02-02T23:39:41
297,385,479
0
1
null
null
null
null
UTF-8
Python
false
false
884
py
import random print(( "+=========================================+\n" "| Guess my number! |\n" "| Please, select a numbers between [0-100]|\n" "| Let's start you have 10 opportunities |\n" "+=========================================+\n" )) my_number = random.randint(0, 100) opportunities = 0 max_opportunities = 10 guess = -1 win = False while not win and opportunities < max_opportunities: guess = int(input('What is the number?: ')) win = (my_number == guess) opportunities += 1 if my_number < guess: print(f'[---] My number is less than {guess}') elif my_number > guess: print(f'[+++] My number is greater than {guess}') else: print('You found my number!') if win: print(f'It only took {opportunities} opportunities, you have won!!!') else: print("Better luck next time")
[ "luis.m.rivas@oracle.com" ]
luis.m.rivas@oracle.com
3972d6d7840538ecae880e9e1fda6b25c54b8bb0
db12b990924703cd74748d8585cd9c11fafa6746
/h2o-py/tests/testdir_apis/H2O_Module/pyunit_h2oas_list.py
3e7ec87a09dad726385a034be07e5ed147db3d49
[ "Apache-2.0" ]
permissive
h2oai/h2o-3
919019a8f297eec676011a9cfd2cc2d97891ce14
d817ab90c8c47f6787604a0b9639b66234158228
refs/heads/master
2023-08-17T18:50:17.732191
2023-08-17T16:44:42
2023-08-17T16:44:42
17,371,412
6,872
2,345
Apache-2.0
2023-09-14T18:05:40
2014-03-03T16:08:07
Jupyter Notebook
UTF-8
Python
false
false
727
py
import sys sys.path.insert(1,"../../../") from tests import pyunit_utils import h2o from h2o.utils.typechecks import assert_is_type def h2oas_list(): """ Python API test: h2o.as_list(data, use_pandas=True, header=True) Copied from pyunit_frame_as_list.py """ iris = h2o.import_file(path=pyunit_utils.locate("smalldata/iris/iris_wheader.csv")) res1 = h2o.as_list(iris, use_pandas=False) assert_is_type(res1, list) res1 = list(zip(*res1)) assert abs(float(res1[0][9]) - 4.4) < 1e-10 and abs(float(res1[1][9]) - 2.9) < 1e-10 and \ abs(float(res1[2][9]) - 1.4) < 1e-10, "incorrect values" if __name__ == "__main__": pyunit_utils.standalone_test(h2oas_list) else: h2oas_list()
[ "noreply@github.com" ]
h2oai.noreply@github.com
c79d336c33025ee87041b0924dfe4af9287a02f3
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p03087/s875530520.py
c6f926255e87de91fe2fee869e4c531b1ffd87e1
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
313
py
N, Q = map(int, input().split()) S = str(input()) li = [] cnt = 0 for i in range(N-1): if S[i]=='A' and S[i+1]=='C': cnt += 1 li.append(cnt) else: li.append(cnt) for i in range(Q): l, r = map(int, input().split()) if l > 1: ans = li[r-2] - li[l-2] else: ans = li[r-2] print(ans)
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
16d806ed7bb7721136f1534dbe98b1aaa13ec985
b15c47a45207e854fb002d69f7e33f8943a5e2b3
/cluster/preprocess/pre_node_merge_text2seq.py
c25a02debf845bcb7cf7835b41e46d424efd847a
[ "Apache-2.0" ]
permissive
yurimkoo/tensormsa
e1af71c00a6b2ec3b3ed35d5adad7bafc34c6fbe
6ad2fbc7384e4dbe7e3e63bdb44c8ce0387f4b7f
refs/heads/master
2021-07-22T13:41:45.110348
2017-11-02T07:13:31
2017-11-02T07:13:31
109,469,204
1
0
null
2017-11-04T05:19:51
2017-11-04T05:19:50
null
UTF-8
Python
false
false
2,165
py
from cluster.preprocess.pre_node import PreProcessNode from master.workflow.preprocess.workflow_pre_merge import WorkFlowPreMerge as WFPreMerge class PreNodeMergeText2Seq(PreProcessNode): """ """ def run(self, conf_data): return True def _init_node_parm(self, key): """ :return: """ wf_conf = WFPreMerge(key) self.batch_size = wf_conf.get_batchsize() self.merge_rule = wf_conf.get_merge_rule() self.merge_type = wf_conf.get_type() self.state_code = wf_conf.get_state_code() def _set_progress_state(self): pass def load_data(self, node_id, parm = 'all'): """ load train data :param node_id: :param parm: :return: """ self._init_node_parm(node_id) if(self.merge_type == 'seq2seq') : return self._merge_seq2seq_type() else : raise Exception ("merge node error: not defined type {0}".format(self.merge_type)) def _merge_seq2seq_type(self): """ merge two data node into one for seq2seq anal :return: """ file_lists = [] encode_data = [] encode_node_list = self.merge_rule['encode_node'] if (len(encode_node_list) > 0): for node_name in encode_node_list: cls_path, cls_name = self.get_cluster_exec_class(str(self.state_code) + "_" + node_name) dyna_cls = self.load_class(cls_path, cls_name) encode_data = encode_data + dyna_cls.load_data(self.state_code + "_" + node_name, parm='all') file_lists.append(encode_data) decode_data = [] decode_node_list = self.merge_rule['decode_node'] if (len(decode_node_list) > 0): for node_name in decode_node_list: cls_path, cls_name = self.get_cluster_exec_class(self.state_code + "_" + node_name) dyna_cls = self.load_class(cls_path, cls_name) decode_data = decode_data + dyna_cls.load_data(self.state_code + "_" + node_name, parm='all') file_lists.append(decode_data) return file_lists
[ "tmddno1@naver.com" ]
tmddno1@naver.com
ae3450610d3f4a3c4469aa2562e125c8a4c0108f
475e2fe71fecddfdc9e4610603b2d94005038e94
/Facebook/moveZeroes.py
f41fef3843fc872dc295c4699bbbe39783028f77
[]
no_license
sidhumeher/PyPractice
770473c699aab9e25ad1f8b7b7cd8ad05991d254
2938c14c2e285af8f02e2cfc7b400ee4f8d4bfe0
refs/heads/master
2021-06-28T20:44:50.328453
2020-12-15T00:51:39
2020-12-15T00:51:39
204,987,730
0
0
null
null
null
null
UTF-8
Python
false
false
572
py
''' Created on Oct 8, 2020 @author: sidteegela ''' ''' Input: [0,1,0,3,12] Output: [1,3,12,0,0] Input: [0,0] Output: [0,0] ''' ''' A two-pointer approach could be helpful here. The idea would be to have one pointer for iterating the array and another pointer that just works on the non-zero elements of the array. ''' def moveZeroes(nums): p = 0 nonZ = 0 while nonZ < len(nums): if nums[nonZ] != 0: nums[p], nums[nonZ] = nums[nonZ], nums[p] p += 1 nonZ += 1 print(nums) if __name__ == '__main__': pass
[ "sidhumeher@yahoo.co.in" ]
sidhumeher@yahoo.co.in
2ff02576828483feb35dafe118dc773c7fb9f7a0
da19363deecd93a73246aaea877ee6607daa6897
/xlsxwriter/test/comparison/test_page_breaks05.py
fb2c64036f0be6ae53d7ae0229ca42202a630f17
[]
no_license
UNPSJB/FarmaciaCrisol
119d2d22417c503d906409a47b9d5abfca1fc119
b2b1223c067a8f8f19019237cbf0e36a27a118a6
refs/heads/master
2021-01-15T22:29:11.943996
2016-02-05T14:30:28
2016-02-05T14:30:28
22,967,417
0
0
null
null
null
null
UTF-8
Python
false
false
1,378
py
############################################################################### # # Tests for XlsxWriter. # # Copyright (c), 2013-2015, John McNamara, jmcnamara@cpan.org # from ..excel_comparsion_test import ExcelComparisonTest from ...workbook import Workbook class TestCompareXLSXFiles(ExcelComparisonTest): """ Test file created by XlsxWriter against a file created by Excel. """ def setUp(self): self.maxDiff = None filename = 'page_breaks05.xlsx' test_dir = 'xlsxwriter/test/comparison/' self.got_filename = test_dir + '_test_' + filename self.exp_filename = test_dir + 'xlsx_files/' + filename self.ignore_files = ['xl/printerSettings/printerSettings1.bin', 'xl/worksheets/_rels/sheet1.xml.rels'] self.ignore_elements = {'[Content_Types].xml': ['<Default Extension="bin"'], 'xl/worksheets/sheet1.xml': ['<pageMargins', '<pageSetup']} def test_create_file(self): """Test the creation of a simple XlsxWriter file with page breaks.""" workbook = Workbook(self.got_filename) worksheet = workbook.add_worksheet() worksheet.set_v_pagebreaks([8, 3, 1, 0]) worksheet.write('A1', 'Foo') workbook.close() self.assertExcelEqual()
[ "lealuque.tw@gmail.com" ]
lealuque.tw@gmail.com
2ae73fa529d6a5d8123828c55ea1d50f63660ebd
4a43cded9d76bd05ca87f037de19ff921a60e151
/13day/07-线程多任务.py
4669c5331f43b6623bc152dc6442a8ad57e8eec5
[]
no_license
lishuang1994/-1807
f7d54c3c93d41539747fc77529d56e73a931676f
23ee6d351a2797d3f5ba7907954d1a93d2745c10
refs/heads/master
2020-03-25T10:09:39.333264
2018-08-22T10:53:49
2018-08-22T10:53:49
143,685,206
0
0
null
null
null
null
UTF-8
Python
false
false
196
py
import time from threading import Thread def saysorry(): print("亲爱的,跪安了") time.sleep(1) for i in range(5): #t = Threadinng(target=saysorry) #t.start() saysorry()
[ "840948928@qq.com" ]
840948928@qq.com
e34d413a8c1c98d6d21b693a46353b5a9c8d1190
163bbb4e0920dedd5941e3edfb2d8706ba75627d
/Code/CodeRecords/2819/60731/291725.py
8bfd0a4aca43e95e2c7b094a9d0e61a6a52acd4a
[]
no_license
AdamZhouSE/pythonHomework
a25c120b03a158d60aaa9fdc5fb203b1bb377a19
ffc5606817a666aa6241cfab27364326f5c066ff
refs/heads/master
2022-11-24T08:05:22.122011
2020-07-28T16:21:24
2020-07-28T16:21:24
259,576,640
2
1
null
null
null
null
UTF-8
Python
false
false
324
py
n=int(input()) data=list(map(int,input().split())) numof1=data.count(1) numof2=data.count(2) numof3=data.count(3) numof4=data.count(4) ans=0 ans+=numof4 num1=int(numof1/4) ans+=num1 numof1-=4*num1 num2=int(numof2/2) ans+=num2 numof2-=2*num2 if (num2+numof3)>=numof1: ans+=(numof2+numof3) else: ans+=numof1 print(ans)
[ "1069583789@qq.com" ]
1069583789@qq.com
55782bca4fc53bfb30a3d26d1336c6538a21a16a
6b8c3974d3ce5f7841e51dcb406666c0c5d92155
/heat/heat_integrationtests/common/clients.py
c7a7f60e6cf45d26c6119542ab40388bb7cfed8a
[ "Apache-2.0" ]
permissive
swjang/cloudexchange
bbbf78a2e7444c1070a55378092c17e8ecb27059
c06ed54f38daeff23166fb0940b27df74c70fc3e
refs/heads/master
2020-12-29T03:18:43.076887
2015-09-21T07:13:22
2015-09-21T07:13:22
42,845,532
1
1
null
2015-09-21T07:13:22
2015-09-21T05:19:35
C++
UTF-8
Python
false
false
5,888
py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import os import ceilometerclient.client import cinderclient.client import heatclient.client import keystoneclient.exceptions import keystoneclient.v2_0.client import neutronclient.v2_0.client import novaclient.client import swiftclient class ClientManager(object): """ Manager that provides access to the official python clients for calling various OpenStack APIs. """ CINDERCLIENT_VERSION = '1' HEATCLIENT_VERSION = '1' NOVACLIENT_VERSION = '2' CEILOMETER_VERSION = '2' def __init__(self, conf): self.conf = conf self.identity_client = self._get_identity_client() self.orchestration_client = self._get_orchestration_client() self.compute_client = self._get_compute_client() self.network_client = self._get_network_client() self.volume_client = self._get_volume_client() self.object_client = self._get_object_client() self.metering_client = self._get_metering_client() def _get_orchestration_client(self): region = self.conf.region endpoint = os.environ.get('HEAT_URL') if os.environ.get('OS_NO_CLIENT_AUTH') == 'True': token = None else: keystone = self._get_identity_client() token = keystone.auth_token try: if endpoint is None: endpoint = keystone.service_catalog.url_for( attr='region', filter_value=region, service_type='orchestration', endpoint_type='publicURL') except keystoneclient.exceptions.EndpointNotFound: return None else: return heatclient.client.Client( self.HEATCLIENT_VERSION, endpoint, token=token, username=self.conf.username, password=self.conf.password) def _get_identity_client(self): return keystoneclient.v2_0.client.Client( username=self.conf.username, password=self.conf.password, tenant_name=self.conf.tenant_name, auth_url=self.conf.auth_url, insecure=self.conf.disable_ssl_certificate_validation) def _get_compute_client(self): dscv = self.conf.disable_ssl_certificate_validation region = self.conf.region client_args = ( self.conf.username, self.conf.password, self.conf.tenant_name, self.conf.auth_url ) # Create our default Nova client to use in testing return novaclient.client.Client( self.NOVACLIENT_VERSION, *client_args, service_type='compute', endpoint_type='publicURL', region_name=region, no_cache=True, insecure=dscv, http_log_debug=True) def _get_network_client(self): auth_url = self.conf.auth_url dscv = self.conf.disable_ssl_certificate_validation return neutronclient.v2_0.client.Client( username=self.conf.username, password=self.conf.password, tenant_name=self.conf.tenant_name, endpoint_type='publicURL', auth_url=auth_url, insecure=dscv) def _get_volume_client(self): auth_url = self.conf.auth_url region = self.conf.region endpoint_type = 'publicURL' dscv = self.conf.disable_ssl_certificate_validation return cinderclient.client.Client( self.CINDERCLIENT_VERSION, self.conf.username, self.conf.password, self.conf.tenant_name, auth_url, region_name=region, endpoint_type=endpoint_type, insecure=dscv, http_log_debug=True) def _get_object_client(self): dscv = self.conf.disable_ssl_certificate_validation args = { 'auth_version': '2.0', 'tenant_name': self.conf.tenant_name, 'user': self.conf.username, 'key': self.conf.password, 'authurl': self.conf.auth_url, 'os_options': {'endpoint_type': 'publicURL'}, 'insecure': dscv, } return swiftclient.client.Connection(**args) def _get_metering_client(self): dscv = self.conf.disable_ssl_certificate_validation keystone = self._get_identity_client() try: endpoint = keystone.service_catalog.url_for( attr='region', filter_value=self.conf.region, service_type='metering', endpoint_type='publicURL') except keystoneclient.exceptions.EndpointNotFound: return None else: args = { 'username': self.conf.username, 'password': self.conf.password, 'tenant_name': self.conf.tenant_name, 'auth_url': self.conf.auth_url, 'insecure': dscv, 'region_name': self.conf.region, 'endpoint_type': 'publicURL', 'service_type': 'metering', } return ceilometerclient.client.Client(self.CEILOMETER_VERSION, endpoint, **args)
[ "kiku4@kinx.net" ]
kiku4@kinx.net
7b56fc22aa32b03fb0cfa39a1d0b49bca82553d5
59d5a801dd8361fe2b68f0cdfc1a0c06bbe9d275
/Competition/fast-flux域名检测/backup/v1/feature_engineering.py
70c7ab417fa009bb50ad40233fda391a9e85cdb7
[]
no_license
HanKin2015/Machine_to_DeepingLearning
2ff377aa68655ca246eb19bea20fec232cec5d77
58fa8d06ef8a8eb0762e7cbd32a09552882c5412
refs/heads/master
2023-01-25T01:16:41.440064
2023-01-18T08:23:49
2023-01-18T08:23:49
134,238,811
0
1
null
null
null
null
UTF-8
Python
false
false
8,016
py
# -*- coding: utf-8 -*- """ 文 件 名: feature_engineering.py 文件描述: 特征工程 作 者: HanKin 创建日期: 2022.10.18 修改日期:2022.10.18 Copyright (c) 2022 HanKin. All rights reserved. """ from common import * def exception_value_processing_by_delete(dataset, feature, lower_threshold, upper_threshold): """异常值处理(删除) """ dataset = dataset[(lower_threshold <= dataset[feature]) & (dataset[feature] <= upper_threshold)] return dataset def exception_value_processing_by_median(dataset, feature, lower_threshold, upper_threshold): """异常值处理(取中位数) """ df = dataset[(lower_threshold <= dataset[feature]) & (dataset[feature] <= upper_threshold)] logger.debug('{}<{},{}>: {}/{}.'.format(feature, lower_threshold, upper_threshold, df.shape[0], dataset.shape[0])) dataset.loc[dataset[feature] < lower_threshold, feature] = df[feature].median() dataset.loc[dataset[feature] > upper_threshold, feature] = df[feature].median() return dataset def exception_value_processing_by_mean(dataset, feature, lower_threshold, upper_threshold): """异常值处理(取平均值) """ df = dataset[(lower_threshold <= dataset[feature]) & (dataset[feature] <= upper_threshold)] logger.debug('{}<{},{}>: {}/{}.'.format(feature, lower_threshold, upper_threshold, df.shape[0], dataset.shape[0])) dataset.loc[dataset[feature] < lower_threshold, feature] = int(df[feature].mean()) dataset.loc[dataset[feature] > upper_threshold, feature] = int(df[feature].mean()) return dataset def missing_value_processing(dataset): """缺失值处理 """ # 用前一个值填充 dataset = dataset.fillna(method='ffill') return dataset def exception_value_processing(dataset): exception_values = [ ['SizeOfStackReserve', 0, 2e7], ['ExportRVA', 0, 2e7], ['DebugRVA', 0, 1e7], ['IATRVA', 0, 1e7], ] for exception_value in exception_values: feature, lower_threshold, upper_threshold = [elem for elem in exception_value] dataset = exception_value_processing_by_mean(dataset, feature, lower_threshold, upper_threshold) if 'label' in dataset.columns: dataset = exception_value_processing_by_delete(dataset, 'SizeOfImage', 1e4, 5e9) return dataset def delete_uncorrelated_features(dataset): """删除相关性低的特征 """ uncorrelated_features = ['time_first', 'time_last', 'rrtype', 'rdata', 'bailiwick'] dataset.drop(uncorrelated_features, axis=1, inplace=True) return dataset def datetime_processing(dataset): """日期时间处理 """ dataset['TimeDateStamp'] = dataset['TimeDateStamp'].apply(lambda x: time.strftime("%Y-%m-%d %X", time.localtime(x))) ts_objs = np.array([pd.Timestamp(item) for item in np.array(dataset['TimeDateStamp'])]) dataset['TS_obj'] = ts_objs # 日期处理(DayName需要转换成数值特征) dataset['Year'] = dataset['TS_obj'].apply(lambda x: x.year) dataset['Month'] = dataset['TS_obj'].apply(lambda x: x.month) dataset['Day'] = dataset['TS_obj'].apply(lambda x: x.day) dataset['DayOfWeek'] = dataset['TS_obj'].apply(lambda x: x.dayofweek) dataset['DayName'] = dataset['TS_obj'].apply(lambda x: x.day_name()) dataset['DayOfYear'] = dataset['TS_obj'].apply(lambda x: x.dayofyear) dataset['WeekOfYear'] = dataset['TS_obj'].apply(lambda x: x.weekofyear) dataset['Quarter'] = dataset['TS_obj'].apply(lambda x: x.quarter) day_name_map = {'Monday': 1, 'Tuesday': 2, 'Wednesday': 3, 'Thursday': 4, 'Friday': 5, 'Saturday': 6, 'Sunday': 7} dataset['DayNameBinMap'] = dataset['DayName'].map(day_name_map) # 时间处理 dataset['Hour'] = dataset['TS_obj'].apply(lambda x: x.hour) dataset['Minute'] = dataset['TS_obj'].apply(lambda x: x.minute) dataset['Second'] = dataset['TS_obj'].apply(lambda x: x.second) #dataset['MUsecond'] = dataset['TS_obj'].apply(lambda x: x.microsecond) #dataset['UTC_offset'] = dataset['TS_obj'].apply(lambda x: x.utcoffset()) ## 按照早晚切分时间 hour_bins = [-1, 5, 11, 16, 21, 23] bin_names = ['LateNight', 'Morning', 'Afternoon', 'Evening', 'Night'] dataset['HourBin'] = pd.cut(dataset['Hour'], bins=hour_bins, labels=bin_names) hour_bin_dummy_features = pd.get_dummies(dataset['HourBin']) dataset = pd.concat([dataset, hour_bin_dummy_features], axis=1) return dataset def discrete_value_processing(dataset): """ """ gle = LabelEncoder() rrname_label = gle.fit_transform(dataset['rrname']) rrname_mapping = {index: label for index, label in enumerate(gle.classes_)} #logger.info(rrname_mapping) dataset['rrname_label'] = rrname_label bailiwick_label = gle.fit_transform(dataset['bailiwick']) bailiwick_mapping = {index: label for index, label in enumerate(gle.classes_)} #logger.info(bailiwick_mapping) dataset['bailiwick_label'] = bailiwick_label dataset['rdata_count'] = dataset['rdata'].apply(lambda x: len(x.split(','))) return dataset def time_processing(dataset): """观察时间处理 """ dataset['time_interval'] = dataset['time_last'] - dataset['time_first'] dataset['time_interval'] = dataset['time_interval'].apply(lambda x: int(x / 86400)) return dataset def features_processing(dataset): # 缺失值处理 #dataset = missing_value_processing(dataset) # 异常值处理 #dataset = exception_value_processing(dataset) # 日期时间处理 #datetime_processing(dataset) time_processing(dataset) # 离散值处理 dataset = discrete_value_processing(dataset) # 删除不相关的特征(相关性低) dataset = delete_uncorrelated_features(dataset) # 特殊处理 return dataset def extended_custom_features(dataset, extended_features_path): """扩展的特征(自定义的字符串特征) """ return dataset def extended_features(dataset, sample_path, extended_features_path): """扩展的特征 """ return dataset def main(): # 获取数据集 train_dataset = pd.read_csv(TRAIN_RAW_DATASET_PATH) train_label = pd.read_csv(TRAIN_LABEL_PATH) test_dataset = pd.read_csv(TEST_RAW_DATASET_PATH) logger.info([train_dataset.shape, train_label.shape, test_dataset.shape]) # 添加标签 train_label.rename(columns = {"domain": "rrname"}, inplace=True) train_dataset = train_dataset.merge(train_label, on='rrname', how='left') logger.info([train_dataset.shape]) # 去除脏数据 # 特征工程 train_dataset = features_processing(train_dataset) test_dataset = features_processing(test_dataset) logger.info('train_dataset: ({}, {}), test_dataset: ({}, {}).'.format( train_dataset.shape[0], train_dataset.shape[1], test_dataset.shape[0], test_dataset.shape[1],)) # 将标签移动到最后一列 label = train_dataset['label'] train_dataset.drop(['label'], axis=1, inplace=True) train_dataset = pd.concat([train_dataset, label], axis=1) # 保存数据集 logger.info([train_dataset.shape, test_dataset.shape]) logger.info(train_dataset.columns) train_dataset.to_csv(TRAIN_DATASET_PATH, sep=',', encoding='utf-8', index=False) test_dataset.to_csv(TEST_DATASET_PATH, sep=',', encoding='utf-8', index=False) if __name__ == '__main__': #os.system('chcp 936 & cls') logger.info('******** starting ********') start_time = time.time() main() end_time = time.time() logger.info('process spend {} s.\n'.format(round(end_time - start_time, 3)))
[ "1058198502@qq.com" ]
1058198502@qq.com
e3c587cf4a2d905939a4c7e8750bc6d0ff07f00c
4c718d78039ca97e1a9b18897642e8bc5cd394d9
/Python code of quizzes/Lesson 6/TagTypes.py
e4bdc42307b8303b26cd8b79cf4aa17d2960ee3f
[]
no_license
aish27/data-wrangle-openstreetmaps-data
81a3fa4a1a06ea8972f6e98af77e9e71350a65bc
9e5eedf49ca67b480bb969ebe4db77df62bf238d
refs/heads/master
2020-12-02T19:26:04.726496
2015-06-12T17:38:37
2015-06-12T17:38:37
37,049,682
0
0
null
null
null
null
UTF-8
Python
false
false
1,562
py
#Determines the problematic tag types that exist in a dataset and their numbers. #!/usr/bin/env python # -*- coding: utf-8 -*- import xml.etree.cElementTree as ET import pprint import re lower = re.compile(r'^([a-z]|_)*$') lower_colon = re.compile(r'^([a-z]|_)*:([a-z]|_)*$') problemchars = re.compile(r'[=\+/&<>;\'"\?%#$@\,\. \t\r\n]') #Uses regular expressions to check each element and find if it contains a problematic tag. def key_type(element, keys): if element.tag == "tag": temp = element.attrib['k'] print temp a = re.search(lower,temp) b = re.search(lower_colon,temp) c = re.search(problemchars,temp) if a!=None: keys["lower"] = keys["lower"] + 1 elif b!=None: keys["lower_colon"] = keys["lower_colon"] + 1 elif c!=None: keys["problemchars"] = keys["problemchars"] + 1 else: keys["other"] = keys["other"] + 1 return keys #Processes the maps and finds probelematic tags. def process_map(filename): keys = {"lower": 0, "lower_colon": 0, "problemchars": 0, "other": 0} for _, element in ET.iterparse(filename): keys = key_type(element, keys) return keys #Uses other methods to find tag types and tests the output. def test(): keys = process_map('example.osm') pprint.pprint(keys) assert keys == {'lower': 5, 'lower_colon': 0, 'other': 1, 'problemchars': 1} if __name__ == "__main__": test()
[ "root@ip-10-47-174-141.ec2.internal" ]
root@ip-10-47-174-141.ec2.internal
a0c7ffbc2b8b323aaf0df3752796182fa2ad9aa5
0d0cf0165ca108e8d94056c2bae5ad07fe9f9377
/12_Dimensionality_Reduction_in_Python/2_Feature_selection_I_selecting_for_feature_information/visualizingTheCorrelationMatrix.py
219bc39cbbc113f94d69e038232a34aa880c624e
[]
no_license
MACHEIKH/Datacamp_Machine_Learning_For_Everyone
550ec4038ebdb69993e16fe22d5136f00101b692
9fe8947f490da221430e6dccce6e2165a42470f3
refs/heads/main
2023-01-22T06:26:15.996504
2020-11-24T11:21:53
2020-11-24T11:21:53
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,245
py
# Visualizing the correlation matrix # Reading the correlation matrix of ansur_df in its raw, numeric format doesn't allow us to get a quick overview. Let's improve this by removing redundant values and visualizing the matrix using seaborn. # Seaborn has been pre-loaded as sns, matplotlib.pyplot as plt, NumPy as np and pandas as pd. # Instructions 1/4 # 100 XP # Create the correlation matrix. # Visualize it using Seaborn's heatmap function. # Instructions 2/4 # 0 XP # Create a boolean mask for the upper triangle of the plot. # Instructions 3/4 # 0 XP # Add the mask to the heatmap. # # Create the correlation matrix (Instruction 1) # corr = ansur_df.corr() # # Draw the heatmap # sns.heatmap(corr, cmap=cmap, center=0, linewidths=1, annot=True, fmt=".2f") # plt.show() # # Create the correlation matrix (Instruction 2) # corr = ansur_df.corr() # # Generate a mask for the upper triangle # mask = np.triu(np.ones_like(corr, dtype=bool)) # Create the correlation matrix (Instruction 3) corr = ansur_df.corr() # Generate a mask for the upper triangle mask = np.triu(np.ones_like(corr, dtype=bool)) # Add the mask to the heatmap sns.heatmap(corr, mask=mask, cmap=cmap, center=0, linewidths=1, annot=True, fmt=".2f") plt.show()
[ "noreply@github.com" ]
MACHEIKH.noreply@github.com
f765a53567eb4d535fe4f1d1fd0c6899ff8f27de
4edbeb3e2d3263897810a358d8c95854a468c3ca
/python3/psutil/count1.py
2930af68edad4e73b64927a1ebaa37e2c82cd7d6
[ "MIT" ]
permissive
jtraver/dev
f505d15d45b67a59d11306cc7252114c265f388b
2197e3443c7619b856470558b737d85fe1f77a5a
refs/heads/master
2023-08-06T02:17:58.601861
2023-08-01T16:58:44
2023-08-01T16:58:44
14,509,952
0
1
MIT
2020-10-14T18:32:48
2013-11-19T00:51:19
Python
UTF-8
Python
false
false
233
py
#!/usr/bin/env python3 #!/usr/bin/python import time import sys def main(): # for count in xrange(10000000): for count in range(10): print("%s" % str(count)) sys.stdout.flush() time.sleep(1) main()
[ "john@aeropsike.com" ]
john@aeropsike.com
22333f24e36f7de94f303b98576d2fb4166f585b
e91fe9c77c39ab855383839820867d5dda27cfdd
/posts/api/pagination.py
feb17a52e7d48c1ab4073025106caa309455469e
[]
no_license
chutchais/Blog-API
6a69580c1f97ecce01e500de56e730864f1e519a
8402d638832703aa4ede7e7e05ac47a902320cd9
refs/heads/master
2021-01-12T03:19:36.874801
2017-01-12T14:48:42
2017-01-12T14:48:42
78,195,920
0
0
null
null
null
null
UTF-8
Python
false
false
249
py
from rest_framework.pagination import ( LimitOffsetPagination, PageNumberPagination, ) class PostLimitOffsetPagination(LimitOffsetPagination): default_limit =2 max_limit=10 class PostPageNumberPagination(PageNumberPagination): page_size=2
[ "chutchai.s@gmail.com" ]
chutchai.s@gmail.com
b878c8557764a724f1d3aeb71971de4a8a664095
2fac796fa58c67fb5a4a95a6e7f28cbef169318b
/python/connecting-graph-ii.py
8dd9c83c7a70a19f748db2b7a41f9c71f45adecf
[]
no_license
jwyx3/practices
f3fe087432e79c8e34f3af3a78dd10278b66dd38
6fec95b9b4d735727160905e754a698513bfb7d8
refs/heads/master
2021-03-12T20:41:59.816448
2019-04-14T06:47:30
2019-04-14T06:47:30
18,814,777
2
0
null
null
null
null
UTF-8
Python
false
false
753
py
class ConnectingGraph2: # @param {int} n def __init__(self, n): self.father = [0 for i in xrange(n + 1)] self.counter = [1 for i in xrange(n + 1)] def find(self, a): if self.father[a] == 0: return a self.father[a] = self.find(self.father[a]) return self.father[a] # @param {int} a, b # return nothing def connect(self, a, b): ra = self.find(a) rb = self.find(b) if ra != rb: self.father[ra] = rb self.counter[rb] += self.counter[ra] # @param {int} a # return {int} the number of nodes connected component # which include a node. def query(self, a): ra = self.find(a) return self.counter[ra]
[ "jwyx88003@gmail.com" ]
jwyx88003@gmail.com
10f6230475b7cd896a058d46d101b08a3ccd9aa9
1885e952aa4a89f8b417b4c2e70b91bf1df887ff
/ABC163/D.py
618f915abc6ca69fb1b9175433ddb970de9c633a
[]
no_license
takumiw/AtCoder
01ed45b4d537a42e1120b1769fe4eff86a8e4406
23b9c89f07db8dd5b5345d7b40a4bae6762b2119
refs/heads/master
2021-07-10T12:01:32.401438
2020-06-27T14:07:17
2020-06-27T14:07:17
158,206,535
1
1
null
null
null
null
UTF-8
Python
false
false
355
py
import sys readline = sys.stdin.buffer.readline MOD = 10 ** 9 + 7 def main(): N, K = map(int, readline().rstrip().split()) ans = 0 m = (N * (N + 1)) // 2 for k in range(K, N+2): mi = ((k-1) * k) // 2 ma = m - ((N-k) * (N-k+1)) // 2 ans += (ma - mi + 1) print(ans % MOD) if __name__ == '__main__': main()
[ "w.tak.1229@gmail.com" ]
w.tak.1229@gmail.com
5ae57a50894ee06d799ee88acb9132651d570cb8
78f54f911d47019da0deeeb6579c7e9e65bb8d21
/src/scheduler/models/dao/common/Log.py
1171ab5340bfce2bfeec5a4abb90b01fc98e85fa
[ "MIT" ]
permissive
jedicontributors/pythondataintegrator
02f8ae1a50cf5ddd85341da738c24aa6a320c442
3e877b367ab9b20185476128ec053db41087879f
refs/heads/main
2023-06-15T07:37:13.313988
2021-07-03T15:46:43
2021-07-03T15:46:43
354,021,102
0
0
MIT
2021-07-03T15:46:44
2021-04-02T13:03:12
Python
UTF-8
Python
false
false
913
py
from sqlalchemy import Column, String, Integer, DateTime from IocManager import IocManager from models.dao.Entity import Entity class Log(Entity, IocManager.Base): __tablename__ = "Log" __table_args__ = {"schema": "Common"} TypeId = Column(Integer, index=True, unique=False, nullable=False) Content = Column(String(4000), index=False, unique=False, nullable=True) LogDatetime = Column(DateTime, index=False, unique=False, nullable=True) JobId = Column(Integer, index=True, unique=False, nullable=True) def __init__(self, TypeId: str = None, Content: str = None, LogDatetime: DateTime = None, JobId: int = None, *args, **kwargs): super().__init__(*args, **kwargs) self.TypeId = TypeId self.Content = Content self.LogDatetime = LogDatetime self.JobId = JobId
[ "ahmetcagriakca@gmail.com" ]
ahmetcagriakca@gmail.com
084706420a235cb05902a3d6090d79a5ce7dd498
3ce592352627591346ea33ea0c2665ad879414e2
/References/opengl/computer.graphics.through.opengl/02_basic_drawing/moveSphere.py
eda2a48c42e3589ed85713e89d5a27c85fe955ba
[ "MIT" ]
permissive
royqh1979/python_libs_usage
113df732ef106f4a5faae1343493756fd703c8c0
57546d5648d8a6b7aca7d7ff9481aa7cd4d8f511
refs/heads/master
2021-04-16T18:14:43.835482
2021-01-11T03:55:25
2021-01-11T03:55:25
249,374,754
0
0
null
null
null
null
UTF-8
Python
false
false
2,672
py
""" // moveSphere.py // // This program allows the user to move a sphere to demonstrate // distortion at the edges of the viewing frustum. // // Interaction: // Press the arrow keys to move the sphere. // Press the space bar to rotate the sphere.. // Press r to reset. // """ from PyQt5 import QtWidgets, QtCore, QtGui, QtOpenGL from OpenGL.GL import * import math as m # Co-ordinates of the sphere center. Xvalue = 0.0 Yvalue = 0.0 Angle = 0.0; # Angle to rotate the sphere. def myWireSphere(r:float, nParal:int, nMerid:int): j=0 while j<m.pi: glBegin(GL_LINE_LOOP) y=r*m.cos(j) i=0 while i < 2*m.pi: x=r*m.cos(i)*m.sin(j) z=r*m.sin(i)*m.sin(j) glVertex3f(x,y,z) i += m.pi / 60 glEnd(); j += m.pi / (nParal + 1) j=0 while j<m.pi: glBegin(GL_LINE_LOOP) i=0 while i < 2*m.pi: x=r*m.sin(i)*m.cos(j) y=r*m.cos(i) z=r*m.sin(j)*m.sin(i) glVertex3f(x,y,z) i += m.pi / 60 glEnd(); j += m.pi / nMerid class MyWidget(QtWidgets.QOpenGLWidget): def __init__(self): super().__init__() self.setFixedSize(500,500) def initializeGL(self) -> None: glClearColor(1.0, 1.0, 1.0, 0.0) def paintGL(self) -> None: glClear(GL_COLOR_BUFFER_BIT) glLoadIdentity() #Set the position of the sphere. glTranslatef(Xvalue, Yvalue, -5.0) glRotatef(Angle, 1.0, 1.0, 1.0) glColor3f(0.0, 0.0, 0.0) myWireSphere(0.5, 16, 10) glFlush() def resizeGL(self, w: int, h: int) -> None: glViewport(0, 0, w, h) glMatrixMode(GL_PROJECTION) glLoadIdentity() glFrustum(-1.0, 1.0, -1.0, 1.0, 1.5, 20.0) glMatrixMode(GL_MODELVIEW) def keyPressEvent(self, event: QtGui.QKeyEvent) -> None: global Xvalue,Yvalue,Angle if event.key() == QtCore.Qt.Key_R: Xvalue = Yvalue = Angle = 0.0 self.update() elif event.key() == QtCore.Qt.Key_Space: Angle += 10.0 self.update() elif event.key() == QtCore.Qt.Key_Up: Yvalue += 0.1 self.update() elif event.key() == QtCore.Qt.Key_Down: Yvalue -= 0.1 self.update() elif event.key() == QtCore.Qt.Key_Left: Xvalue -= 0.1 self.update() elif event.key() == QtCore.Qt.Key_Right: Xvalue += 0.1 self.update() if __name__ == '__main__': app = QtWidgets.QApplication([]) win = MyWidget() win.show() app.exec()
[ "royqh1979@gmail.com" ]
royqh1979@gmail.com
e0287539bd801f859c461d1ae1d70b8055288938
5759c0ed3219c06437ce5b39ef9ad92b5e191fed
/py/0428_serialize_and_deserialize_n_ary_tree.py
1df779c0ca32628d1c3067e97cc5b065abbbda52
[]
no_license
mengnan1994/Surrender-to-Reality
ba69df7c36112ad19f19157a9f368eae6340630f
66232728ce49149188f863271ec2c57e426abb43
refs/heads/master
2022-02-25T01:34:49.526517
2019-09-22T17:21:28
2019-09-22T17:21:28
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,773
py
""" Serialization is the process of converting a data structure or object into a sequence of bits so that it can be stored in a file or memory buffer, or transmitted across a network connection link to be reconstructed later in the same or another computer environment. Design an algorithm to serialize and deserialize an N-ary tree. An N-ary tree is a rooted tree in which each node has no more than N children. There is no restriction on how your serialization/deserialization algorithm should work. You just need to ensure that an N-ary tree can be serialized to a string and this string can be deserialized to the original tree structure. For example, you may serialize the following 3-ary tree as [1 [3[5 6] 2 4]]. You do not necessarily need to follow this format, so please be creative and come up with different approaches yourself. Note: 1. N is in the range of [1, 1000] 2. Do not use class member/global/static variables to store states. Your serialize and deserialize algorithms should be stateless. """ class Node(object): def __init__(self, val, children): self.val = val self.children = children class Codec(object): def serialize(self, root): serialization = str(self.inorder_helper(root)) print(serialization) res = "" for symbol in serialization: if symbol != " " and symbol != ",": res += symbol return res def inorder_helper(self, node): if not node.children: return [node.val, []] return [node.val] + [self.inorder_helper(child) for child in node.children] def deserialize(self, data): if not data: return None scanner = 0 dummy_head = Node("dummy head", []) stack = [dummy_head] val_stack = ["#"] root = None while scanner < len(data): # print(val_stack) # print(data[scanner:]) # print("===") if data[scanner] == '[': # 如果遇到左括号,准备入栈 scanner += 1 if data[scanner] == ']': scanner += 1 elif data[scanner] in ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9']: val = "" while data[scanner] in ['0', '1', '2', '3', '4', '5', '6', '7', '8', '9']: val += data[scanner] scanner += 1 val = int(val) node = Node(val, []) stack[-1].children.append(node) stack.append(node) val_stack.append(node.val) else: root = stack.pop() val_stack.pop() scanner += 1 return root
[ "hurricane.cui@gmail.com" ]
hurricane.cui@gmail.com
18f43854543ea9c9c8c46129f56e665c7042d7d6
de24f83a5e3768a2638ebcf13cbe717e75740168
/moodledata/vpl_data/468/usersdata/304/111685/submittedfiles/Av2_Parte3.py
8f2fd9356f208463f08f81498b9a718876b02d30
[]
no_license
rafaelperazzo/programacao-web
95643423a35c44613b0f64bed05bd34780fe2436
170dd5440afb9ee68a973f3de13a99aa4c735d79
refs/heads/master
2021-01-12T14:06:25.773146
2017-12-22T16:05:45
2017-12-22T16:05:45
69,566,344
0
0
null
null
null
null
UTF-8
Python
false
false
390
py
# -*- coding: utf-8 -*- m = int(input('Quantas listas: ')) n = int(input('Qntd elemento listas: ')) for i in range (0,m,1): lista=[] for j in range (0,n,1): lista.append(int(input('Elemento: '))) media = sum(lista)/len(lista) soma = 0 for i in range (0,(n-1),1): soma += ((i - media)) dp = ((1/(n-1))*(soma)**2)**(1/2) print('%.2f'%media) print('%.2f'%dp)
[ "rafael.mota@ufca.edu.br" ]
rafael.mota@ufca.edu.br
8e858adedb2bd3195fa4dfb7bb2b4c6af2189603
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p03910/s110121278.py
4c9ecfd6bfbe6621fcbe2a07ea7568326440d590
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
153
py
N = int(input()) s = 0 j = 1 for i in range(1, N+1): s += i if s >= N: j = i break k = s - N for i in range(1, j+1): if i == k: continue print(i)
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
fa1fa3378a017fe7890883215c6f4ed40f43828a
a2b696ba8edd5e6b8aa1a4c4aea19cc56e6beb66
/api/serializers.py
d4db895abe2ac6146ca319be0a8d40433d07655a
[]
no_license
kyrios213/drf-django
7163eeaba2f32cd0f1cfa5871718c1892a7dc83a
d61438bbb919e7ab845da14bf04cfd85d108de73
refs/heads/main
2023-06-05T03:19:33.440410
2021-06-21T03:23:46
2021-06-21T03:23:46
378,791,890
0
0
null
null
null
null
UTF-8
Python
false
false
428
py
from rest_framework import serializers, viewsets from django.contrib.auth import get_user_model class UserSerializer(serializers.ModelSerializer): class Meta: model = get_user_model() fields = ['id', 'email', 'password'] extra_kwargs = {'password': {'write_only': True, 'min_length': 8}} def create(self, validated_data): return get_user_model().objects.create_user(**validated_data)
[ "kyrios213@gmail.com" ]
kyrios213@gmail.com
9132e1ad54fca3fe47ba425f3bbb3e1377737f65
6b6bf72694e5aa6425f11c956e4a5371b2c73e09
/populate/main.py
bcccf70630692e0c3d23db7eb0615a5c25787907
[]
no_license
danielspeixoto/elasticsearch-workshop
5d9dccebb80bf58b786f167ff8eb6de8a0282a8c
9872831361ea349453b8e202eb02880c43cbc238
refs/heads/master
2020-04-02T01:18:28.858625
2018-10-25T22:24:28
2018-10-25T22:24:28
153,848,788
0
0
null
null
null
null
UTF-8
Python
false
false
252
py
from helpers.Index import Index, connect from helpers.XMLRepository import XMLRepository connection = connect("localhost", "9200") index = Index(connection, "meu-index") dados = XMLRepository("./datasets/bitcoin/") index.bulk_insert(dados.data())
[ "danielspeixoto@outlook.com" ]
danielspeixoto@outlook.com
15aa14bd147b72e801f55c46ff70efe32d7ef80e
d6074aac6e9e5f2fa5355c3c9ddaebe892c4151d
/setup.py
8813b411347ea2ca52a5f6274befae3a458b8877
[ "BSD-2-Clause" ]
permissive
pombredanne/pydeps
f5aae8b825cf86db8308adad808b4bbf8cee52a2
95833f42f86f9c5c23ef14b675828d0ddc4df73e
refs/heads/master
2021-01-18T07:52:41.522028
2016-05-19T13:53:14
2016-05-19T13:53:14
null
0
0
null
null
null
null
UTF-8
Python
false
false
913
py
import setuptools from distutils.core import setup setup( name='pydeps', version='1.2.7', packages=['pydeps'], install_requires=[ 'enum34' ], long_description=open('README.rst').read(), entry_points={ 'console_scripts': [ 'pydeps = pydeps.pydeps:pydeps', ] }, url='https://github.com/thebjorn/pydeps', license='BSD', author='bjorn', author_email='bp@datakortet.no', description='Display module dependencies', classifiers=[ 'Intended Audience :: Developers', 'Natural Language :: English', 'License :: OSI Approved :: BSD License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules', ] )
[ "bp@datakortet.no" ]
bp@datakortet.no
7cb57f1aee5711236c63768e429265d71ba59e80
21b461b71b4c63f7aac341bd12ba35d211c7956e
/codes/03_func/Graph_Permutation_isomorphic/01_test_permutation.py
587c4d9c282ed70920f61d81d9c4546575e05260
[]
no_license
Ziaeemehr/workshop_scripting
cebdcb552720f31fd6524fd43f257ca46baf70e2
ed5f232f6737bc9f750d704455442f239d4f0561
refs/heads/main
2023-08-22T23:00:36.121267
2023-07-19T10:53:41
2023-07-19T10:53:41
153,342,386
4
0
null
null
null
null
UTF-8
Python
false
false
771
py
""" A_G = P A_H P.T P is permutation matrix P.T is the transpose matrix of T """ import numpy as np import networkx as nx def permutation(alpha, n): P = np.zeros((n, n), dtype=int) for i in range(n): P[i][alpha[i]] = 1 return P A_G = np.array([[0, 1, 0, 0, 0], [1, 0, 1, 0, 1], [0, 1, 0, 1, 1], [0, 0, 1, 0, 1], [0, 1, 1, 1, 0]]) A_H = np.array([[0, 1, 1, 1, 0], [1, 0, 1, 0, 0], [1, 1, 0, 1, 0], [1, 0, 1, 0, 1], [0, 0, 0, 1, 0]]) # arbitrary permutation of nodes. alpha = [4, 3, 0, 1, 2] P = permutation(alpha, 5) A_Gp = np.matmul(np.matmul(P, A_H), P.T) # check if it works? print(np.array_equal(A_Gp, A_G))
[ "a.ziaeemehr@gmail.com" ]
a.ziaeemehr@gmail.com
38bbf0422041425be9f4f1aac557fcfbfe23a739
694d57c3e512ce916269411b51adef23532420cd
/leetcode_review2/74search_a_2D_matrix.py
866fe5944bf1f8a5a7c20dee6738ef99ba24a70b
[]
no_license
clovery410/mycode
5541c3a99962d7949832a0859f18819f118edfba
e12025e754547d18d5bb50a9dbe5e725fd03fd9c
refs/heads/master
2021-05-16T02:46:47.996748
2017-05-10T23:43:50
2017-05-10T23:43:50
39,235,141
1
1
null
null
null
null
UTF-8
Python
false
false
421
py
class Solution(object): def searchMatrix(self, matrix, target): m, n = len(matrix), len(matrix[0]) if len(matrix) > 0 else 0 cur_i, cur_j = 0, n-1 while cur_i < m and cur_j >= 0: if target == matrix[cur_i][cur_j]: return True if target < matrix[cur_i][cur_j]: cur_j -= 1 else: cur_i += 1 return False
[ "seasoul410@gmail.com" ]
seasoul410@gmail.com
b76b50f294cf1d7560a3699d3704b41173c7ea8c
c956401119e44e41f3873b4734c857eda957e2cd
/metrics/lr_smape.py
8a71c4a237fb069f751070fdc252cea4f1b598f7
[]
no_license
zhekunz2/c4pp
6f689bce42215507d749b8b2be96c7f68ed8c49c
9f8054addb48d9440662d8c8f494359846423ffd
refs/heads/master
2020-09-24T06:17:29.866975
2019-12-03T17:41:25
2019-12-03T17:41:25
225,678,894
0
0
null
null
null
null
UTF-8
Python
false
false
880
py
#!/usr/bin/env python import os import json import sys import numpy as np try: smape_thres=float(sys.argv[1]) weight=float(sys.argv[2]) bias=float(sys.argv[3]) data_dir=sys.argv[4] with open(os.path.join(data_dir, 'data.json')) as json_data: d = json.load(json_data) datas = [x for x in d.keys() if x=='x' or x=='y'] data_x = sorted(datas)[0] data_y = sorted(datas)[1] predict = [x * float(weight) + float(bias) for x in d[data_x]] observe = d[data_y] ape = np.zeros(len(observe)) for i,p in enumerate(predict): o = observe[i] ape[i] = abs(o - p) / ((abs(o) + abs(p))) smape = np.mean(ape) if abs(smape) <= smape_thres: print('True: ' + str(smape)) else: print('False: ' + str(smape)) except: print('Error') exit(0)
[ "zhekunz2@Zhekuns-MacBook-Pro.local" ]
zhekunz2@Zhekuns-MacBook-Pro.local
7a144841bd9fe8ec5e8663356c33fac86b2dbf31
acb082b215e6d214a5065f76c0454dcf1fb2a533
/src/cobra/core/auth/access.py
e3881977e6cf7cd7420cdef66f2da916e6890e6d
[ "Apache-2.0" ]
permissive
lyoniionly/django-cobra
130b25cd897cc94b6a8da722e9a83ecea3b00c49
2427e5cf74b7739115b1224da3306986b3ee345c
refs/heads/master
2016-09-06T10:23:10.607025
2016-02-22T14:16:49
2016-02-22T14:16:49
29,646,851
1
0
null
null
null
null
UTF-8
Python
false
false
3,147
py
from __future__ import absolute_import from cobra.core.loading import get_model __all__ = ['from_user', 'from_member', 'DEFAULT', 'SCOPES'] OrganizationMember = get_model('organization', 'OrganizationMember') SCOPES = set([ 'org:read', 'org:write', 'org:delete', 'team:read', 'team:write', 'team:delete', 'project:read', 'project:write', 'project:delete', 'event:read', 'event:write', 'event:delete', ]) class BaseAccess(object): is_global = False is_active = False sso_is_valid = False teams = () scopes = frozenset() def has_scope(self, scope): if not self.is_active: return False return scope in self.scopes def has_team(self, team): if not self.is_active: return False if self.is_global: return True return team in self.teams class Access(BaseAccess): # TODO(dcramer): this is still a little gross, and ideally backend access # would be based on the same scopes as API access so theres clarity in # what things mean def __init__(self, scopes, is_global, is_active, teams, sso_is_valid): self.teams = teams self.scopes = scopes self.is_global = is_global self.is_active = is_active self.sso_is_valid = sso_is_valid def from_user(user, organization): if user.is_superuser: return Access( scopes=SCOPES, is_global=True, is_active=True, teams=(), sso_is_valid=True, ) if not organization: return DEFAULT try: om = OrganizationMember.objects.get( user=user, organization=organization, ) except OrganizationMember.DoesNotExist: return DEFAULT return from_member(om) def from_member(member): # TODO(dcramer): we want to optimize this access pattern as its several # network hops and needed in a lot of places if member.has_global_access: teams = () else: teams = member.teams.all() # try: # auth_provider = AuthProvider.objects.get( # organization=member.organization_id, # ) # except AuthProvider.DoesNotExist: # sso_is_valid = True # else: # try: # auth_identity = AuthIdentity.objects.get( # auth_provider=auth_provider, # ) # except AuthIdentity.DoesNotExist: # sso_is_valid = False # else: # sso_is_valid = auth_identity.is_valid(member) sso_is_valid = True return Access( is_global=member.has_global_access, is_active=True, sso_is_valid=sso_is_valid, scopes=member.scopes, teams=teams, ) class NoAccess(BaseAccess): @property def sso_is_valid(self): return True @property def is_global(self): return False @property def is_active(self): return False @property def teams(self): return () @property def scopes(self): return frozenset() DEFAULT = NoAccess()
[ "beihua_445566@126.com" ]
beihua_445566@126.com
59620cdbfa9167aeb555cca708e1f7ede5412081
7d31324f874130bc5059314048193f474f2a820a
/gui.py
6a41ea6e86f10145d7f19d02e451b0a0f571e17f
[]
no_license
deshudiosh/PyTSF
2ae13c9724b6e290016aad6329db5c175bd319fa
6c3a8b6b35a4fc602f10aa10aff72d9bc887ccdb
refs/heads/master
2021-01-20T01:36:11.138441
2017-10-27T18:57:31
2017-10-27T18:57:31
89,306,166
0
0
null
null
null
null
UTF-8
Python
false
false
1,025
py
import remi.gui as gui from remi import start, App class PyTsfGui(App): def __init__(self, *args): super(PyTsfGui, self).__init__(*args) def main(self): # margin 0px auto allows to center the app to the screen container = gui.VBox(width=400, margin='0px auto') container.style['background'] = '#808080' logo = gui.Label('PyTSF', width='80%', height=60, margin='0px auto') logo.style['margin'] = 'auto' panel = gui.HBox(width=400, height=100, margin='0px auto') dropdown = gui.DropDown() refresh = gui.Button('R') options = gui.Button("O") go = gui.Button("Go!") panel.append(dropdown) panel.append(refresh) panel.append(options) panel.append(go) container.append(logo) container.append(panel) # returning the root widget return container def startApp(): start(PyTsfGui, address="127.0.0.20", debug=True, enable_file_cache=False, multiple_instance=True)
[ "deshudiosh@gmail.com" ]
deshudiosh@gmail.com
2bc3743b880efe5c8a8739a55179572019820af5
c31e46965ea47cdb0c61a6b525aecea45dbf4d0b
/gram/views.py
b442f39e96c2ad10dff14a8042bb03981ca1e2de
[]
no_license
marysinaida/-Instagram
687563c7b9a44adcdd09217ed02ff5c2c24623d2
cb0768ad24b7650db6b7fbfd9d445232b154cece
refs/heads/master
2022-12-06T07:02:55.786066
2020-01-08T06:57:47
2020-01-08T06:57:47
228,771,439
0
0
null
2022-11-22T05:13:53
2019-12-18T06:16:03
Python
UTF-8
Python
false
false
2,960
py
from django.shortcuts import render,get_object_or_404 from django.utils import timezone from .forms import PostForm from .models import Post,Profile from django.contrib.auth.models import User from django.views.generic import (ListView,CreateView,DetailView) from django.http import HttpResponse from django.contrib.auth.decorators import login_required # Create your views here. class PostListView(ListView): template_name = "post_list.html" queryset = Post.objects.all().filter(created_date__lte=timezone.now()).order_by('-created_date') context_object_name = "posts" success_url = '/' class PostCreateView(CreateView): template_name = "post_create.html" form_class = PostForm queryset = Post.objects.all() success_url = '/' def form_valid(self,form): print(form.cleaned_data) form.instance.author = self.request.user return super().form_valid(form) class PostDetailView(DetailView): template_name ="post_details.html" queryset = Post.objects.all().filter(created_date__lte=timezone.now()) def get_object(self): id_ = self.kwargs.get('id') return get_object_or_404(Post,id=id_) def signUp(request): return render(request,'registration/registration_form.html') @login_required(login_url='/accounts/login/') def login(request): return render(request,'registration/login.html') @login_required(login_url='/profile') def search_results(request): if 'username' in request.GET and request.GET["username"]: search_term = request.GET.get("username") searched_users = User.objects.filter(username = search_term) message = f"{search_term}" profile_pic = User.objects.all() return render(request, 'search.html', {'message':message, 'results':searched_users, 'profile_pic':profile_pic}) else: message = "You haven't searched for any term" return render(request, 'search.html', {'message':message}) def profile(request): # image = request.user.profile.posts.all() if request.method == 'POST': user_form = UpdateUserForm(request.POST, instance=request.user) prof_form = UpdateUserProfileForm(request.POST, request.FILES, instance=request.user.profile) if user_form.is_valid() and prof_form.is_valid(): user_form.save() prof_form.save() return HttpResponseRedirect(request.path_info) return render(request, 'profile.html', {}) return render(request, 'profile.html', {}) def timeline(request): posts = Post.objects.all().filter(created_date__lte=timezone.now()).order_by('-created_date') if request.method == 'POST': form = PostForm(request.POST, request.FILES) if edit_form.is_valid(): form.save() return render(request, 'post_list.html', {'form': form, 'posts': posts}) return render(request, 'post_list.html', {'posts': posts})
[ "marydorcassinaida54@gmail.com" ]
marydorcassinaida54@gmail.com
e0d376cdb3b6cbbbbeea5b2ea6c2a81b88fbf34c
422c9cc1c5ef7eba24610e66d6a74ec2e16bf39e
/install_isolated/lib/python2.7/dist-packages/turtle_actionlib/msg/_ShapeFeedback.py
332cdc621b2cf527210a07a88ad56ccb86dca474
[]
no_license
twighk/ROS-Pi3
222c735d3252d6fce43b427cdea3132f93025002
9f2912c44ae996040f143c1e77e6c714162fc7d2
refs/heads/master
2021-01-01T05:16:20.278770
2016-05-08T19:24:15
2016-05-08T19:24:15
58,306,257
2
0
null
null
null
null
UTF-8
Python
false
false
2,927
py
# This Python file uses the following encoding: utf-8 """autogenerated by genpy from turtle_actionlib/ShapeFeedback.msg. Do not edit.""" import sys python3 = True if sys.hexversion > 0x03000000 else False import genpy import struct class ShapeFeedback(genpy.Message): _md5sum = "d41d8cd98f00b204e9800998ecf8427e" _type = "turtle_actionlib/ShapeFeedback" _has_header = False #flag to mark the presence of a Header object _full_text = """# ====== DO NOT MODIFY! AUTOGENERATED FROM AN ACTION DEFINITION ====== #feedback """ __slots__ = [] _slot_types = [] def __init__(self, *args, **kwds): """ Constructor. Any message fields that are implicitly/explicitly set to None will be assigned a default value. The recommend use is keyword arguments as this is more robust to future message changes. You cannot mix in-order arguments and keyword arguments. The available fields are: :param args: complete set of field values, in .msg order :param kwds: use keyword arguments corresponding to message field names to set specific fields. """ if args or kwds: super(ShapeFeedback, self).__init__(*args, **kwds) def _get_types(self): """ internal API method """ return self._slot_types def serialize(self, buff): """ serialize message into buffer :param buff: buffer, ``StringIO`` """ try: pass except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize(self, str): """ unpack serialized message in str into this message instance :param str: byte array of serialized message, ``str`` """ try: end = 0 return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill def serialize_numpy(self, buff, numpy): """ serialize message with numpy array types into buffer :param buff: buffer, ``StringIO`` :param numpy: numpy python module """ try: pass except struct.error as se: self._check_types(struct.error("%s: '%s' when writing '%s'" % (type(se), str(se), str(locals().get('_x', self))))) except TypeError as te: self._check_types(ValueError("%s: '%s' when writing '%s'" % (type(te), str(te), str(locals().get('_x', self))))) def deserialize_numpy(self, str, numpy): """ unpack serialized message in str into this message instance using numpy for array types :param str: byte array of serialized message, ``str`` :param numpy: numpy python module """ try: end = 0 return self except struct.error as e: raise genpy.DeserializationError(e) #most likely buffer underfill _struct_I = genpy.struct_I
[ "twighk@outlook.com" ]
twighk@outlook.com
36da4a1a8ee737c91a2e4cfd313a97f8c36a836a
56482e0b2ce6517fff41d0f78e0c0ed000d977a1
/fmcapi/api_objects/helper_functions.py
73723b685d3181277032d2a1b3b56c4c54e1fb90
[ "BSD-3-Clause" ]
permissive
banzigaga/fmcapi
ab4d7aaaf4be4f2b0686d07b6272f8b9531577da
fd924de96e200ca8e0d5088b27a5abaf6f915bc6
refs/heads/master
2020-12-11T14:45:07.896571
2019-12-12T20:02:07
2019-12-12T20:02:07
233,876,405
1
0
BSD-3-Clause
2020-01-14T15:46:26
2020-01-14T15:46:26
null
UTF-8
Python
false
false
4,316
py
"""Misc methods/functions that are used by the fmcapi package's modules.""" import re import ipaddress import json import logging logging.debug(f"In the {__name__} module.") def syntax_correcter(value, permitted_syntax="""[.\w\d_\-]""", replacer="_"): """ Check 'value' for invalid characters (identified by 'permitted_syntax') and replace them with 'replacer'. :param value: (str) String to be checked. :param permitted_syntax: (str) regex of allowed characters. :param replacer: (str) character used to replace invalid characters. :return: (str) Modified string with "updated" characters. """ logging.debug("In syntax_correcter() helper_function.") new_value = "" for char in range(0, len(value)): if not re.match(permitted_syntax, value[char]): new_value += replacer else: new_value += value[char] return new_value def get_networkaddress_type(value): """ Check to see whether 'value' is a host, range, or network. :param value: (str) x.x.x.x, x.x.x.x/xx, or x.x.x.x-x.x.x.x :return: (str) 'host', 'network', or 'range' """ logging.debug("In get_networkaddress_type() helper_function.") if "/" in value: ip, bitmask = value.split("/") if ip == "32" or bitmask == "128": return "host" else: return "network" else: if "-" in value: return "range" else: return "host" def is_ip(ip): """ Check whether the provided string is an IP address. :param ip: (str) x.x.x.x :return: (boolean) """ logging.debug("In is_ip() helper_function.") try: ipaddress.ip_address(ip) except ValueError as err: logging.error(err) return False return True def is_ip_network(ip): """ Check whether provided string is a valid network address. See if the provided IP/SM is the "network address" of the subnet. :param ip: (str) x.x.x.x/xx :return: (boolean) """ logging.debug("In is_ip_network() helper_function.") try: ipaddress.ip_network(ip) except ValueError as err: logging.error(err) return False return True def validate_ip_bitmask_range(value="", value_type=""): """ We need to check the provided IP address (or range of addresses) and make sure the IPs are valid. :param value: (str) x.x.x.x, x.x.x.x/xx, or x.x.x.x-x.x.x.x :param value_type: (str) 'host', 'network', or 'range' :return: (dict) {value=value_fixed, valid=boolean} """ logging.debug("In validate_ip_bitmask_range() helper_function.") return_dict = {"value": value, "valid": False} if value_type == "range": for ip in value.split("-"): if is_ip(ip): return_dict["valid"] = True elif value_type == "host" or value_type == "network": if is_ip_network(value): return_dict["valid"] = True return return_dict["valid"] def mocked_requests_get(**kwargs): """ Use to "mock up" a response from using the "requests" library to avoid actually using the "requests" library. :param kwargs: :return: (boolean) """ logging.debug("In mocked_requests_get() helper_function.") class MockResponse: def __init__(self, **kwargs): logging.debug("In MockResponse __init__ method.") self.text = json.dumps(kwargs["text"]) self.status_code = kwargs["status_code"] def close(self): logging.debug("In MockResponse close method.") return True return MockResponse(**kwargs) def validate_vlans(start_vlan, end_vlan=""): """ Validate that the start_vlan and end_vlan numbers are in 1 - 4094 range. If not, then return 1, 4094. :param start_vlan: (int) Lower VLAN number in range. :param end_vlan: (int) Upper VLAN number in range. :return: (int) start_vlan, (int) end_vlan) """ logging.debug("In validate_vlans() helper_function.") if end_vlan == "": end_vlan = start_vlan if int(end_vlan) < int(start_vlan): start_vlan, end_vlan = end_vlan, start_vlan if 0 < int(start_vlan) < 4095 and 0 < int(end_vlan) < 4095: return start_vlan, end_vlan else: return 1, 4094
[ "dmickels@cisco.com" ]
dmickels@cisco.com
7036e36ad6e119dac97c2175715c6b24857595cf
5a52ccea88f90dd4f1acc2819997fce0dd5ffb7d
/alipay/aop/api/response/AnttechOceanbaseVerificationcodeVerifyResponse.py
2808330b22ff50c4c507563278735baa72cdba13
[ "Apache-2.0" ]
permissive
alipay/alipay-sdk-python-all
8bd20882852ffeb70a6e929038bf88ff1d1eff1c
1fad300587c9e7e099747305ba9077d4cd7afde9
refs/heads/master
2023-08-27T21:35:01.778771
2023-08-23T07:12:26
2023-08-23T07:12:26
133,338,689
247
70
Apache-2.0
2023-04-25T04:54:02
2018-05-14T09:40:54
Python
UTF-8
Python
false
false
479
py
#!/usr/bin/env python # -*- coding: utf-8 -*- import json from alipay.aop.api.response.AlipayResponse import AlipayResponse class AnttechOceanbaseVerificationcodeVerifyResponse(AlipayResponse): def __init__(self): super(AnttechOceanbaseVerificationcodeVerifyResponse, self).__init__() def parse_response_content(self, response_content): response = super(AnttechOceanbaseVerificationcodeVerifyResponse, self).parse_response_content(response_content)
[ "jishupei.jsp@alibaba-inc.com" ]
jishupei.jsp@alibaba-inc.com
f5c468beecaa5a2685e310d47134afa02ab63714
386d5d4f8f102e701d02b326cd066f520e3dff9f
/ProjectApplication/project_core/migrations/0166_organisation_long_name_english.py
3e520cccaf379fa51b06988bbdeee96f5063231b
[ "MIT" ]
permissive
Swiss-Polar-Institute/project-application
ae2561c3ae2c1d5412d165d959ce2e5886135e0a
7dc4a9f7e0f8d28c89977b85f99bc5e35ea77d43
refs/heads/master
2023-08-31T04:01:23.492272
2023-08-25T14:33:02
2023-08-25T14:33:02
206,330,401
7
5
MIT
2023-09-13T08:03:53
2019-09-04T13:49:39
Python
UTF-8
Python
false
false
496
py
# Generated by Django 3.2.3 on 2021-05-24 14:47 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('project_core', '0165_callcareerstage_enabled'), ] operations = [ migrations.AddField( model_name='organisation', name='long_name_english', field=models.CharField(blank=True, help_text='English name by which the organisation is known', max_length=100, null=True), ), ]
[ "jenny_t152@yahoo.co.uk" ]
jenny_t152@yahoo.co.uk
09499b5a40b99525ef5398364dfabc50873929aa
d69995905dcd2522e53082c32c10f582b6779cba
/apps/testsuites/utils.py
046c1755163e1f7524679acbd2cd94c4a979ed31
[]
no_license
liqi629/learn_nm_drf
010b35ab4254267a601a13f29e2ea3adb3be615e
ad0eef6fe2b338d613974850977de51b9d82ccc6
refs/heads/master
2023-02-10T14:33:33.521450
2021-01-13T09:14:23
2021-01-13T09:14:23
324,942,039
0
1
null
null
null
null
UTF-8
Python
false
false
854
py
import re from datetime import datetime from apps.testcases.models import Testcases def modify_output(results): datas_list = [] for item in results: mtch = re.search(r'(.*)T(.*)\..*?',item['create_time']) item['create_time'] = mtch.group(1) +' '+mtch.group(2) mtch = re.search(r'(.*)T(.*)\..*?', item['update_time']) item['update_time'] = mtch.group(1) + ' ' + mtch.group(2) datas_list.append(item) return datas_list def get_testcases_by_interface_ids(ids_list): """ 通过接口ID获取用例 :param ids_list: :return: """ one_list =[] for interface_id in ids_list: testcases_qs = Testcases.objects.values_list('id',flat=True).\ filter(interface_id=interface_id,is_delete=False) one_list.extend(list(testcases_qs)) return one_list
[ "liqi_629@163.com" ]
liqi_629@163.com
cb5c4ccba717cd9c2920942e071419e95cf3aa0d
2d276785c3663d4798be462115291c4706dbd255
/Python从菜鸟到高手/chapter2/demo2.03.py
5e0bf41280f733cc7bad2b80a3e3d347a440b9a8
[]
no_license
bupthl/Python
81c92433bd955663e6cda5fe7cab5ea3d067c3de
bdb33aeeb179a43100b9ef7129a925c63a133fd3
refs/heads/master
2022-02-21T11:02:40.195265
2019-08-16T05:49:18
2019-08-16T05:49:18
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,049
py
''' --------《Python从菜鸟到高手》源代码------------ 欧瑞科技版权所有 作者:李宁 如有任何技术问题,请加QQ技术讨论群:264268059 或关注“极客起源”订阅号或“欧瑞科技”服务号或扫码关注订阅号和服务号,二维码在源代码根目录 如果QQ群已满,请访问https://geekori.com,在右侧查看最新的QQ群,同时可以扫码关注公众号 “欧瑞学院”是欧瑞科技旗下在线IT教育学院,包含大量IT前沿视频课程, 请访问http://geekori.com/edu或关注前面提到的订阅号和服务号,进入移动版的欧瑞学院 “极客题库”是欧瑞科技旗下在线题库,请扫描源代码根目录中的小程序码安装“极客题库”小程序 关于更多信息,请访问下面的页面 https://geekori.com/help/videocourse/readme.html ''' # 需要在Windows下执行 import os import sys f_handler=open('out.log', 'w') oldstdout = sys.stdout sys.stdout=f_handler os.system('cls') sys.stdout = oldstdout
[ "registercn@outlook.com" ]
registercn@outlook.com
90f5d0ecdb0f132e76809d77dca276ad9a766253
f576f0ea3725d54bd2551883901b25b863fe6688
/sdk/eventhub/azure-eventhub-checkpointstoretable/azure/eventhub/extensions/checkpointstoretable/_vendor/data/tables/aio/_base_client_async.py
06b3136394d128cf3c39652676f430c56596b3be
[ "LicenseRef-scancode-generic-cla", "MIT", "LGPL-2.1-or-later" ]
permissive
Azure/azure-sdk-for-python
02e3838e53a33d8ba27e9bcc22bd84e790e4ca7c
c2ca191e736bb06bfbbbc9493e8325763ba990bb
refs/heads/main
2023-09-06T09:30:13.135012
2023-09-06T01:08:06
2023-09-06T01:08:06
4,127,088
4,046
2,755
MIT
2023-09-14T21:48:49
2012-04-24T16:46:12
Python
UTF-8
Python
false
false
6,981
py
# ------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for # license information. # -------------------------------------------------------------------------- from typing import Any, List, Mapping, Optional, Union, TYPE_CHECKING from uuid import uuid4 from azure.core.credentials import AzureSasCredential, AzureNamedKeyCredential from azure.core.pipeline.policies import ( ContentDecodePolicy, AsyncBearerTokenCredentialPolicy, AsyncRedirectPolicy, DistributedTracingPolicy, HttpLoggingPolicy, UserAgentPolicy, ProxyPolicy, AzureSasCredentialPolicy, RequestIdPolicy, CustomHookPolicy, NetworkTraceLoggingPolicy, ) from azure.core.pipeline.transport import ( AsyncHttpTransport, HttpRequest, ) from .._generated.aio import AzureTable from .._base_client import AccountHostsMixin, get_api_version, extract_batch_part_metadata from .._authentication import SharedKeyCredentialPolicy from .._constants import STORAGE_OAUTH_SCOPE from .._error import RequestTooLargeError, TableTransactionError, _decode_error from .._policies import StorageHosts, StorageHeadersPolicy from .._sdk_moniker import SDK_MONIKER from ._policies_async import AsyncTablesRetryPolicy if TYPE_CHECKING: from azure.core.credentials_async import AsyncTokenCredential class AsyncTablesBaseClient(AccountHostsMixin): def __init__( # pylint: disable=missing-client-constructor-parameter-credential self, endpoint: str, *, credential: Optional[Union[AzureSasCredential, AzureNamedKeyCredential, "AsyncTokenCredential"]] = None, **kwargs: Any ) -> None: super(AsyncTablesBaseClient, self).__init__(endpoint, credential=credential, **kwargs) # type: ignore self._client = AzureTable( self.url, policies=kwargs.pop('policies', self._policies), **kwargs ) self._client._config.version = get_api_version(kwargs, self._client._config.version) # pylint: disable=protected-access async def __aenter__(self): await self._client.__aenter__() return self async def __aexit__(self, *args): await self._client.__aexit__(*args) async def close(self) -> None: """This method is to close the sockets opened by the client. It need not be used when using with a context manager. """ await self._client.close() def _configure_credential(self, credential): # type: (Any) -> None if hasattr(credential, "get_token"): self._credential_policy = AsyncBearerTokenCredentialPolicy( # type: ignore credential, STORAGE_OAUTH_SCOPE ) elif isinstance(credential, SharedKeyCredentialPolicy): self._credential_policy = credential # type: ignore elif isinstance(credential, AzureSasCredential): self._credential_policy = AzureSasCredentialPolicy(credential) # type: ignore elif isinstance(credential, AzureNamedKeyCredential): self._credential_policy = SharedKeyCredentialPolicy(credential) # type: ignore elif credential is not None: raise TypeError("Unsupported credential: {}".format(credential)) def _configure_policies(self, **kwargs): return [ RequestIdPolicy(**kwargs), StorageHeadersPolicy(**kwargs), UserAgentPolicy(sdk_moniker=SDK_MONIKER, **kwargs), ProxyPolicy(**kwargs), self._credential_policy, ContentDecodePolicy(response_encoding="utf-8"), AsyncRedirectPolicy(**kwargs), StorageHosts(**kwargs), AsyncTablesRetryPolicy(**kwargs), CustomHookPolicy(**kwargs), NetworkTraceLoggingPolicy(**kwargs), DistributedTracingPolicy(**kwargs), HttpLoggingPolicy(**kwargs), ] async def _batch_send(self, *reqs: "HttpRequest", **kwargs) -> List[Mapping[str, Any]]: """Given a series of request, do a Storage batch call.""" # Pop it here, so requests doesn't feel bad about additional kwarg policies = [StorageHeadersPolicy()] changeset = HttpRequest("POST", None) # type: ignore changeset.set_multipart_mixed( *reqs, policies=policies, boundary="changeset_{}".format(uuid4()) ) request = self._client._client.post( # pylint: disable=protected-access url="https://{}/$batch".format(self._primary_hostname), headers={ "x-ms-version": self.api_version, "DataServiceVersion": "3.0", "MaxDataServiceVersion": "3.0;NetFx", "Content-Type": "application/json", "Accept": "application/json" }, ) request.set_multipart_mixed( changeset, policies=policies, enforce_https=False, boundary="batch_{}".format(uuid4()), ) pipeline_response = await self._client._client._pipeline.run(request, **kwargs) # pylint: disable=protected-access response = pipeline_response.http_response # TODO: Check for proper error model deserialization if response.status_code == 413: raise _decode_error( response, error_message="The transaction request was too large", error_type=RequestTooLargeError) if response.status_code != 202: raise _decode_error(response) parts_iter = response.parts() parts = [] async for p in parts_iter: parts.append(p) error_parts = [p for p in parts if not 200 <= p.status_code < 300] if any(error_parts): if error_parts[0].status_code == 413: raise _decode_error( response, error_message="The transaction request was too large", error_type=RequestTooLargeError) raise _decode_error( response=error_parts[0], error_type=TableTransactionError, ) return [extract_batch_part_metadata(p) for p in parts] class AsyncTransportWrapper(AsyncHttpTransport): """Wrapper class that ensures that an inner client created by a `get_client` method does not close the outer transport for the parent when used in a context manager. """ def __init__(self, async_transport): self._transport = async_transport async def send(self, request, **kwargs): return await self._transport.send(request, **kwargs) async def open(self): pass async def close(self): pass async def __aenter__(self): pass async def __aexit__(self, *args): # pylint: disable=arguments-differ pass
[ "noreply@github.com" ]
Azure.noreply@github.com
feb64c06c982a14449c43ee53cc8a6183106fe90
093b7b3ce929fa1383d5c1571271807f171aec23
/rocket/entities/api/endpoint.py
971b53d78b7948b93078e3f8e4b2b12806ebba05
[]
no_license
takaaki-mizuno/smart-rocket-cli
fa728a13bc80781aea182922f7a0ee2c546518b1
1b4a522c54e94857803f9be246f0c480e50dae3d
refs/heads/master
2020-03-21T02:06:54.115723
2018-06-25T22:43:32
2018-06-25T22:43:32
137,980,655
0
0
null
null
null
null
UTF-8
Python
false
false
934
py
from .parameter import Parameter from .response import Reference class Endpoint: def __init__(self, path, method, definition, spec): self.path = path self.method = method self.definition = definition self.spec = spec self.parameters = [] self.response = None self.parse() def parse(self): self.parse_parameters() self.parse_response() def parse_parameters(self): if 'parameters' not in self.definition: return for parameter in self.definition['parameters']: self.parameters.append(Parameter(parameter)) def parse_response(self): if 'responses' not in self.definition: return for name, definition in self.definition['responses'].items(): if name[0:1] == '2' and 'schema' in definition: self.response = Reference(definition['schema']['$ref'])
[ "takaaki.mizuno@gmail.com" ]
takaaki.mizuno@gmail.com
8d5c12675fbee82c4d20320e48951b0b9f1842e2
f716ec8240b775170283cb6d43da50d0ff3561b7
/testkraken/testing_functions/check_output.py
9eeb9c524b75d5c5bbe91ba9471dee63bc318ab0
[]
no_license
jdkent/testkraken
63493df89a5ae427dab9da14fc794de96cda6cef
3b5d83f0eeaad7969902cfa867e6b4a1d4d7e691
refs/heads/master
2020-06-01T11:39:57.286709
2019-06-07T09:55:16
2019-06-07T09:55:16
190,766,540
0
0
null
2019-06-07T15:24:49
2019-06-07T15:24:49
null
UTF-8
Python
false
false
3,766
py
#/usr/bin/env python from __future__ import division import json import os, inspect from glob import glob import pandas as pd import numpy as np import pdb def creating_dataframe(files_list): """ reads every json file from the files_list and creates one data frame """ outputmap = {0: 'voxels', 1: 'volume'} df = pd.DataFrame() for (i, filename) in enumerate(files_list): with open(filename, 'rt') as fp: in_dict = json.load(fp) # in cwl i'm loosing the directory name #subject = filename.split(os.path.sep)[-3] subject = "subject_{}".format(i) in_dict_mod = {} for k, v in in_dict.items(): if isinstance(v, list): for idx, value in enumerate(v): in_dict_mod["%s_%s" % (k, outputmap[idx])] = value else: in_dict_mod[k] = v df[subject] = pd.Series(in_dict_mod) return df.T def check_output(file_out, file_ref=None, name=None, **kwargs): if type(file_ref) is list: expected_files = file_ref elif type(file_ref) is str: expected_files = [file_ref] if type(file_out) is list: output_files = file_out elif type(file_out) is str: output_files = [file_out] df_exp = creating_dataframe(expected_files) df_out = creating_dataframe(output_files) #df_exp.to_csv('output/ExpectedOutput.csv') #df_out.to_csv('output/ActualOutput.csv') # DJ TOD: this doesn't work, check with the original repo #df_diff = df_exp - df_out #df_diff = df_diff.dropna() report_filename = "report_{}.json".format(name) out = {} # chosing just a few columns keys_test = ["white_voxels", "gray_voxels", "csf_voxels", "Right-Hippocampus_voxels", "Right-Amygdala_voxels", "Right-Caudate_voxels"] out["index_name"] = list(df_exp.index) for key in keys_test: out["re_{}".format(key.replace("_voxels", "").replace("Right-", "R-"))] = [] for subj in df_exp.index: for key in keys_test: if df_exp.loc[subj, key] != 0.: out["re_{}".format(key.replace("_voxels", "").replace("Right-", "R-"))].append(round( 1. * abs(df_exp.loc[subj, key] - df_out.loc[subj, key]) / df_exp.loc[subj, key], 5)) elif df_out.loc[subj, key] != 0.: out["re_{}".format(key.replace("_voxels", "").replace("Right-", "R-"))].append(1.) else: out["re_{}".format(key.replace("_voxels", "").replace("Right-", "R-"))].append(0.) out["regr"] = [] for i, subj in enumerate(out["index_name"]): list_tmp = [] for k in out.keys(): if k not in ["index_name", "regr"]: list_tmp.append(out[k][i]) try: assert max(list_tmp) < 0.05 out["regr"].append("PASSED") except(AssertionError): out["regr"].append("FAILED") #out_max = {"max_diff": max(diff)} with open(report_filename, "w") as f: json.dump(out, f) if __name__ == "__main__": from argparse import ArgumentParser, RawTextHelpFormatter defstr = ' (default %(default)s)' parser = ArgumentParser(description=__doc__, formatter_class=RawTextHelpFormatter) parser.add_argument("-out", nargs="+", dest="file_out", help="file with the output for testing") parser.add_argument("-ref", nargs="+", dest="file_ref", help="file with the reference output") parser.add_argument("-name", dest="name", help="name of the test provided by a user") args = parser.parse_args() check_output(**vars(args))
[ "djarecka@gmail.com" ]
djarecka@gmail.com
533ca7f861f7a6a34ff0265d601d017f0f9835f7
b6d82335dfe93f86977e4cbafe592eff32536712
/src/aws_hashicorp_packer_reaper/schema.py
f028da830fb0a93c2da236b6aabc65f3a6ed1a36
[ "Apache-2.0" ]
permissive
felixubierar/aws-hashicorp-packer-reaper
2bb9709b3c95c53d519fa5d009cbafc48232b1b4
9b29f6896f43a01f47d2a894059e4afa191a8ff2
refs/heads/master
2023-05-19T17:01:23.872830
2021-06-14T08:57:50
2021-06-14T08:57:50
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,835
py
import durations from aws_hashicorp_packer_reaper.logger import log from jsonschema.exceptions import ValidationError from jsonschema import validators, Draft7Validator, FormatChecker, validators schema = { "$schema": "http://json-schema.org/draft-04/schema#", "type": "object", "properties": { "mode": { "type": "string", "description": "of operations", "enum": ["stop", "terminate"], }, "older_than": { "type": "string", "description": "period since launched", "format": "duration", "default": "2h", }, "dry_run": { "type": "boolean", "description": "if you only want output", "default": False, }, "tags": { "type": "array", "description": "to select EC2 instances with", "items": {"type": "string", "minLength": 1}, }, }, "required": [ "mode", "older_than", ], } @FormatChecker.cls_checks("duration") def duration_checker(value) -> bool: """ checks whether the `value` is a valid duration. >>> duration_checker({}) False >>> duration_checker(1.0) False >>> duration_checker("2h") True >>> duration_checker("hundred days") False """ try: if isinstance(value, str): durations.Duration(value) return True except durations.exceptions.InvalidTokenError as e: pass return False def extend_with_default(validator_class): validate_properties = validator_class.VALIDATORS["properties"] def set_defaults(validator, properties, instance, schema): for property, subschema in properties.items(): if "default" in subschema: instance.setdefault(property, subschema["default"]) for error in validate_properties( validator, properties, instance, schema, ): yield error return validators.extend( validator_class, {"properties": set_defaults}, ) validator = extend_with_default(Draft7Validator)(schema, format_checker=FormatChecker()) def validate(request: dict) -> bool: """ return True and completes the missing values if the dictionary matches the schema, otherwise False. >>> validate({"mode": "stoep"}) False >>> validate({"mode": "stop", "older_than": "sdfsdfsf dagen"}) False >>> x = {"mode": "stop"} >>> validate(x) True >>> print(x) {'mode': 'stop', 'older_than': '2h', 'dry_run': False} """ try: validator.validate(request) return True except ValidationError as e: log.error("invalid request received: %s" % str(e.message)) return False
[ "mark.van.holsteijn@gmail.com" ]
mark.van.holsteijn@gmail.com
3d6fae9b8527d4f8823a630b37b8fecb7d2d8207
b9f21bc90eed396dde950c30a1b482be0fb8ba30
/AtCoder/ABC/114-B_1.py
85c142e193b383c2ee95f438be3c37a3edec97b1
[]
no_license
nanigasi-san/nanigasi
127a21db1b31759908fd74cebabe240e5abf8267
5e3c3e78344dd9558cafe439beb272b9a80d0f3a
refs/heads/master
2020-04-03T18:57:40.132489
2019-06-19T15:03:38
2019-06-19T15:03:38
155,504,101
1
0
null
2019-02-20T09:40:05
2018-10-31T05:33:11
Python
UTF-8
Python
false
false
155
py
S = list(input()) list = [] for _ in range(len(S)-2): Sx = S[_]+S[_+1]+S[_+2] sa = int(Sx)-753 list.append(abs(sa)) list.sort() print(list[0])
[ "nanigasi.py@gmail.com" ]
nanigasi.py@gmail.com
a4f92fc07660a377a8f5ca8b05dec21c348b7a8e
2caa47f0bdb2f03469a847c3ba39496de315d992
/Contest/ABC086/c/main.py
1e87c7d1f6b031f11c2f42d5104c63354529a69a
[ "CC0-1.0" ]
permissive
mpses/AtCoder
9023e44885dc67c4131762281193c24b69d3b6da
9c101fcc0a1394754fcf2385af54b05c30a5ae2a
refs/heads/master
2023-03-23T17:00:11.646508
2021-03-20T12:21:19
2021-03-20T12:21:19
287,489,233
0
0
null
null
null
null
UTF-8
Python
false
false
233
py
#!/usr/bin/env python3 n, *a = map(int, open(0).read().split()) a = [0]*3 + a for i in range(n): i *= 3 t, x, y = map(lambda j:abs(a[i+3+j]-a[i+j]), [0,1,2]) d = x+y if d>t or d%2-t%2: print("No"); exit() print("Yes")
[ "nsorangepv@gmail.com" ]
nsorangepv@gmail.com
671f43d908baa3c118b7e3e44c09aee84552f4d5
ff58ba25d940ed34d9684efab04adef85d1e1c0f
/ENV/lib/python2.6/site-packages/gunicorn/app/pasterapp.py
388478181468378e0cb86f5300a8946ea4738159
[]
no_license
afsmith/Kneto-Sello
e9046a81ff83652531adc55aab3f90f77af5b5be
a1b12daf8a04ef485ddcaa2944b2d87878a8cdd0
refs/heads/master
2021-03-27T17:31:23.830989
2013-06-04T07:29:58
2013-06-04T07:29:58
6,720,999
0
0
null
null
null
null
UTF-8
Python
false
false
5,270
py
# -*- coding: utf-8 - # # This file is part of gunicorn released under the MIT license. # See the NOTICE for more information. import logging import os import pkg_resources import sys import ConfigParser from paste.deploy import loadapp, loadwsgi SERVER = loadwsgi.SERVER from gunicorn.app.base import Application from gunicorn.config import Config class PasterBaseApplication(Application): def app_config(self): cx = loadwsgi.loadcontext(SERVER, self.cfgurl, relative_to=self.relpath) gc, lc = cx.global_conf.copy(), cx.local_conf.copy() cfg = {} host, port = lc.pop('host', ''), lc.pop('port', '') if host and port: cfg['bind'] = '%s:%s' % (host, port) elif host: cfg['bind'] = host cfg['workers'] = int(lc.get('workers', 1)) cfg['umask'] = int(lc.get('umask', 0)) cfg['default_proc_name'] = gc.get('__file__') for k, v in gc.items(): if k not in self.cfg.settings: continue cfg[k] = v for k, v in lc.items(): if k not in self.cfg.settings: continue cfg[k] = v return cfg def configure_logging(self): if hasattr(self, "cfgfname"): self.logger = logging.getLogger('gunicorn') # from paste.script.command parser = ConfigParser.ConfigParser() parser.read([self.cfgfname]) if parser.has_section('loggers'): if sys.version_info >= (2, 6): from logging.config import fileConfig else: # Use our custom fileConfig -- 2.5.1's with a custom Formatter class # and less strict whitespace (which were incorporated into 2.6's) from gunicorn.logging_config import fileConfig config_file = os.path.abspath(self.cfgfname) fileConfig(config_file, dict(__file__=config_file, here=os.path.dirname(config_file))) return super(PasterBaseApplication, self).configure_logging() class PasterApplication(PasterBaseApplication): def init(self, parser, opts, args): if len(args) != 1: parser.error("No application name specified.") cfgfname = os.path.normpath(os.path.join(os.getcwd(), args[0])) cfgfname = os.path.abspath(cfgfname) if not os.path.exists(cfgfname): parser.error("Config file not found: %s" % cfgfname) self.cfgurl = 'config:%s' % cfgfname self.relpath = os.path.dirname(cfgfname) self.cfgfname = cfgfname sys.path.insert(0, self.relpath) pkg_resources.working_set.add_entry(self.relpath) return self.app_config() def load(self): return loadapp(self.cfgurl, relative_to=self.relpath) class PasterServerApplication(PasterBaseApplication): def __init__(self, app, gcfg=None, host="127.0.0.1", port=None, *args, **kwargs): self.cfg = Config() self.app = app self.callable = None gcfg = gcfg or {} cfgfname = gcfg.get("__file__") if cfgfname is not None: self.cfgurl = 'config:%s' % cfgfname self.relpath = os.path.dirname(cfgfname) self.cfgfname = cfgfname cfg = kwargs.copy() if port and not host.startswith("unix:"): bind = "%s:%s" % (host, port) else: bind = host cfg["bind"] = bind if gcfg: for k, v in list(gcfg.items()): cfg[k] = v cfg["default_proc_name"] = cfg['__file__'] try: for k, v in list(cfg.items()): if k.lower() in self.cfg.settings and v is not None: self.cfg.set(k.lower(), v) except Exception, e: sys.stderr.write("\nConfig error: %s\n" % str(e)) sys.stderr.flush() sys.exit(1) self.configure_logging() def load_config(self): if not hasattr(self, "cfgfname"): return cfg = self.app_config() for k,v in cfg.items(): try: self.cfg.set(k.lower(), v) except: sys.stderr.write("Invalid value for %s: %s\n\n" % (k, v)) raise def load(self): if hasattr(self, "cfgfname"): return loadapp(self.cfgurl, relative_to=self.relpath) return self.app def run(): """\ The ``gunicorn_paster`` command for launcing Paster compatible apllications like Pylons or Turbogears2 """ from gunicorn.app.pasterapp import PasterApplication PasterApplication("%prog [OPTIONS] pasteconfig.ini").run() def paste_server(app, gcfg=None, host="127.0.0.1", port=None, *args, **kwargs): """\ A paster server. Then entry point in your paster ini file should looks like this: [server:main] use = egg:gunicorn#main host = 127.0.0.1 port = 5000 """ from gunicorn.app.pasterapp import PasterServerApplication PasterServerApplication(app, gcfg=gcfg, host=host, port=port, *args, **kwargs).run()
[ "andrew.smith@kneto.fi" ]
andrew.smith@kneto.fi
1382dcbea411dd5c7a5227efcd1b1545775e46c7
8ecf97e0f12037ccd0b63f265fddd3bf84229d7a
/keras/inception_autoencoder_predict.py
966fa0964b70a98e8e0ee7e382925390c8b2781a
[]
no_license
ArifSohaib/video_collission_detection
158af02ce55d8e7f39b532417010b882106bdb2f
79ac145f4a3d0386ffac4f4fc2cc6efda056e53f
refs/heads/master
2021-09-11T01:36:09.085256
2018-04-05T19:44:40
2018-04-05T19:44:40
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,739
py
from inception_autoencoder import build_autoencoder import numpy as np import pickle import matplotlib.pyplot as plt from sklearn import preprocessing def main(): impact_data = np.loadtxt('../data/features/impacts_period1.npy') full_data = np.loadtxt('../data/features/period1_full.npy') min_max_scaler = preprocessing.MinMaxScaler() impact_data = min_max_scaler.fit_transform(impact_data) full_data = min_max_scaler.fit_transform(full_data) model = build_autoencoder() model.load_weights('../data/weights/autoencoder_weights.h5') predict_impact = model.predict(impact_data) predict_full = model.predict(full_data) mse_full = ((predict_full - full_data) ** 2).mean(axis=1) mse_impact = ((predict_impact - impact_data) ** 2).mean(axis=1) mean_mse = mse_full.mean() print("full mse avg {}".format(mean_mse)) print("impact mse avg {}".format(mse_impact.mean())) print("full mse min {}".format(mse_full.min())) print("impact mse min {}".format(mse_impact.min())) print("full mse max {}".format(mse_full.max())) print("impact mse max {}".format(mse_impact.max())) plt.hist(mse_full, label='full_mse stats') plt.show() plt.hist(mse_impact, label='impact_mse stats') plt.show() full_percentile = np.percentile(mse_full, 50) impact_percentile = np.percentile(mse_impact, 50); print("full mse percentile {}".format(full_percentile)) print("impact mse percentile {}".format(impact_percentile)) print("length of full data {}".format(len(mse_full))) pred_impact_idx = [] #running the above statistics, we can say that if the mse is above the max of impact mse, then it is not an impact for idx, err in enumerate(mse_full): if err > impact_percentile: pred_impact_idx.append(idx) with open('../data/frames/frames_vid1_div5.pkl', 'rb') as f: confirmed_idx = pickle.load(f) confirmed_idx = sorted(confirmed_idx.values()) """ for each value in confirmed_idx we need 10 numbers before and 10 after it(totaling 29 * 20 = 580) """ full_idx = [] for idx in confirmed_idx: for i in range(-10, 10): full_idx.append(idx+i) true_count = 0 false_pos = 0 #to check accuracy, we can compare against idx's computed before idx_count = 0; for idx in pred_impact_idx: if idx in full_idx: true_count += 1 else: false_pos += 1 print("num predictions {}".format(len(pred_impact_idx))) print("true count {}".format(true_count)) print("length of pred impacts {}".format(len(full_idx))) print("false pos {}".format(false_pos)) if __name__ == '__main__': main()
[ "arif_sohaib@outlook.com" ]
arif_sohaib@outlook.com
490e204bc88e69d4ea255b9c9b4e172ee01ae582
eb26f2a53339cc9880c193225919bd37078113aa
/flask/mysite/app.py
1ce6098bfcafb1745e57a8640c7e316044ddd05d
[]
no_license
nnyong/TIL-c9
c2590ea94c13221a45f274beb22a0f03b09bf6d4
2e2803ee60467ffbad7d9a704b0029476ebead0b
refs/heads/master
2020-04-17T17:17:53.798516
2019-05-02T03:08:56
2019-05-02T03:08:56
166,777,119
0
0
null
null
null
null
UTF-8
Python
false
false
1,614
py
import os, csv from flask import Flask, render_template, request app=Flask(__name__) @app.route('/') def index(): return 'Hello World!' @app.route('/greeting/<string:name>') def greeting(name): return f'반갑습니다! {name}님!' @app.route('/cube/<int:num>') def cube(num): result=num**3 return str(result) @app.route('/html_file') def html_file(): return render_template('html_file.html') @app.route('/hi/<name>') def hi(name): return render_template('hi.html', name_in_html=name) @app.route('/fruits') def fruits(): fruits=['apple','banana','mango','melon'] return render_template('fruits.html',fruits=fruits) @app.route('/send') def send(): return render_template('send.html') @app.route('/receive') def receive(): # request.args # {'who':'junwoo','message':'hello'} who=request.args.get('who') message=request.args.get('message') with open('guestbook.csv','a',encoding='utf8',newline='') as f: writer=csv.DictWriter(f,fieldnames=['who','message']) writer.writerow({ 'who': who, 'message': message }) return render_template('receive.html',name=who, message=message) @app.route('/guestbook') def guestbook(): messages=[] with open('guestbook.csv','r',encoding='utf8',newline='') as f: reader=csv.DictReader(f) for row in reader: messages.append(row) return render_template('guestbook.html',messages=messages) if __name__=='__main__': app.run(host=os.getenv('IP'), port=os.getenv('PORT'), debug=True)
[ "jo.youthful@gmail.com" ]
jo.youthful@gmail.com
e9ec52b3a17bacb2dda1181371225ad96a07f778
90e4fe85b70221ae82f99ca6930da980ef8f674a
/Anul_1/Semestru_1/FundamentalsOfProg/Seminar/s08p1_I6/src/store/ui/console.py
e9bdf474e451dc086c98f9ea7956991ddff0ca45
[]
no_license
stefangeorgescu970/university-assignments
108235e047b963efb6cd1b952f6b96849e1dc3d3
9253cc084b74a62035c96a4a2accfbba43812c16
refs/heads/master
2022-12-10T14:49:20.299356
2020-04-28T14:05:41
2020-04-28T14:05:41
259,648,446
0
0
null
2022-12-07T20:33:05
2020-04-28T13:43:02
C++
UTF-8
Python
false
false
2,116
py
""" @author: radu """ import traceback from store.domain.dto import OrderDTO from store.domain.validators import StoreException from util.common import MyUtil class Console(object): def __init__(self, product_controller, order_controller, statistics_controller): self.__product_controller = product_controller self.__order_controller = order_controller self.__statistics_controller = statistics_controller def run_console(self): # TODO implement an menu or cmd based console self.__init_data() print("all products:") self.__print_all_products() print("all orders:") self.__print_all_orders() print("products filtered by name (name containing the string 'p'):") MyUtil.print_list(self.__product_controller.filter_products_by_name("p")) print("the cost of all orders is: ", self.__statistics_controller.compute_all_orders_cost()) print("the orders with the cost greater than 2 is:") MyUtil.print_list(self.__statistics_controller.filter_orders(2)) self.__print_sorted_orders() def __print_all_products(self): MyUtil.print_list(self.__product_controller.get_all()) def __print_all_orders(self): MyUtil.print_list(self.__order_controller.get_all()) def __init_data(self): try: self.__product_controller.add_product(1, "p1", 100) self.__product_controller.add_product(2, "p2", 200) self.__product_controller.add_product(3, "bla", 300) self.__order_controller.add_order(1, 1, 2) self.__order_controller.add_order(2, 1, 3) self.__order_controller.add_order(3, 2, 4) except StoreException as se: print("exception when initializing data: ", se) traceback.print_exc() def __print_sorted_orders(self): print("the orders sorted descending by cost and ascending by name:") sorted_orders = self.__statistics_controller.sort_orders() for i in range(0, len(sorted_orders)): print(i + 1, sorted_orders[i])
[ "stefan.georgescu.970@gmail.com" ]
stefan.georgescu.970@gmail.com
b65f85e32d7186281a2604cc160bd674c86d13a5
7aecab27c231c5207f26a1682543b0d6c5093c06
/server/dancedeets/util/korean_dates.py
c6c4fe2fe48b14f79ae353c0f2e1153c24e04d9e
[]
no_license
mikelambert/dancedeets-monorepo
685ed9a0258ea2f9439ae4ed47ebf68bb5f89256
4eff1034b9afd3417d168750ea3acfaecd20adc6
refs/heads/master
2022-08-10T07:16:32.427913
2018-04-15T22:05:58
2018-04-15T22:05:58
75,126,334
24
2
null
2022-07-29T22:28:45
2016-11-29T22:04:44
Python
UTF-8
Python
false
false
2,382
py
# -*-*- encoding: utf-8 -*-*- import datetime import re DAWN = 6 # sun rises at 6am-ish _DATETIME_SPAN_SEPARATOR = ur'(?:~|/|부터)' # '부터' = 'from' _D_DATE = ur'(?P<day>\d+)[일\(]' _MD_DATE = ur'(?:(?P<month>\d+)월\s*)?' + _D_DATE _YMD_DATE = ur'(?:(?P<year>\d+)[년녀]\s*)?' + _MD_DATE # WEEKDAY = r'(?:\(.\)| ..일)?' _AM_PM_TIME = ur'(?:(?P<ampm>오후|오전) )?(?P<hour>\d+)시 ?(?:(?P<minute>\d+)분|(?P<half>반))?' _TIME = ur'(?:(?P<dawn>새벽)|%s)' % _AM_PM_TIME def _extract_date(m, date_default=None): return datetime.date( int(m.group('year') or date_default.year), int(m.group('month') or date_default.month), int(m.group('day') or date_default.day) ) def _extract_time(m, time_default=None): if m.group('dawn'): return datetime.time(DAWN) if m.group('half'): minute = 30 else: minute = int(m.group('minute') or 0) if unicode(m.group('ampm')) == u'오후': ampm_offset = 12 elif m.group('ampm') == u'오전': ampm_offset = 0 else: ampm_offset = 12 hour = int(m.group('hour')) if hour == 12: hour = 0 return datetime.time(hour + ampm_offset, minute) def parse_times(s): elems = re.split(_DATETIME_SPAN_SEPARATOR, s, 2) if len(elems) == 1: start_str, end_str = elems[0], None else: start_str, end_str = elems[0], elems[1] start_date_match = re.search(_YMD_DATE, start_str) start_time_match = re.search(_TIME, start_str) start_datetime = _extract_date(start_date_match) if start_time_match: start_datetime = datetime.datetime.combine(start_datetime, _extract_time(start_time_match)) end_datetime = None if end_str: end_date_match = re.search(_YMD_DATE, end_str) end_time_match = re.search(_TIME, end_str) if end_date_match or end_time_match: if end_date_match: end_datetime = _extract_date(end_date_match, date_default=start_datetime) else: if isinstance(start_datetime, datetime.datetime): end_datetime = start_datetime.date() else: end_datetime = start_datetime if end_time_match: end_datetime = datetime.datetime.combine(end_datetime, _extract_time(end_time_match)) return (start_datetime, end_datetime)
[ "mlambert@gmail.com" ]
mlambert@gmail.com
220ac4e18bfaea6b2a5fc122ae8bcc38836508db
fafb89a3552e4dbb47d134966462ef5f3f37f576
/KEMP/v0.6_ovelap_ok/fdtd3d/cpu/test/pbc_multi_plot.py
218f06c00053697b3fa19d4b987ec597ccaa7b61
[]
no_license
EMinsight/fdtd_accelerate
78fa1546df5264550d12fba3cf964838b560711d
a566c60753932eeb646c4a3dea7ed25c7b059256
refs/heads/master
2021-12-14T03:26:52.070069
2012-07-25T08:25:21
2012-07-25T08:25:21
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,528
py
import numpy as np import sys, os sys.path.append( os.path.expanduser('~') ) from kemp.fdtd3d.cpu import QueueTask, Fields, Core, Pbc, IncidentDirect tmax = 150 tfunc = lambda tstep: np.sin(0.05 * tstep) # plot import matplotlib.pyplot as plt import matplotlib as mpl mpl.rc('image', interpolation='nearest', origin='lower') fig = plt.figure(figsize=(14,8)) # z-axis nx, ny, nz = 180, 160, 2 fields = Fields(QueueTask(), nx, ny, nz) Core(fields) Pbc(fields, 'xyz') IncidentDirect(fields, 'ey', (20, 0, 0), (20, -1, -1), tfunc) IncidentDirect(fields, 'ex', (0, 20, 0), (-1, 20, -1), tfunc) for tstep in xrange(1, tmax+1): fields.update_e() fields.update_h() fields.enqueue_barrier() ax1 = fig.add_subplot(2, 3, 1) ax1.imshow(fields.get('ey')[:,:,nz/2].T, vmin=-1.1, vmax=1.1) ax1.set_title('%s, ey[20,:,:]' % repr(fields.ns)) ax1.set_xlabel('x') ax1.set_ylabel('y') ax2 = fig.add_subplot(2, 3, 4) ax2.imshow(fields.get('ex')[:,:,nz/2].T, vmin=-1.1, vmax=1.1) ax2.set_title('%s, ex[:,20,:]' % repr(fields.ns)) ax2.set_xlabel('x') ax2.set_ylabel('y') # y-axis nx, ny, nz = 180, 2, 160 fields = Fields(QueueTask(), nx, ny, nz) Core(fields) Pbc(fields, 'xyz') IncidentDirect(fields, 'ez', (20, 0, 0), (20, -1, -1), tfunc) IncidentDirect(fields, 'ex', (0, 0, 20), (-1, -1, 20), tfunc) for tstep in xrange(1, tmax+1): fields.update_e() fields.update_h() fields.enqueue_barrier() ax1 = fig.add_subplot(2, 3, 2) ax1.imshow(fields.get('ez')[:,ny/2,:].T, vmin=-1.1, vmax=1.1) ax1.set_title('%s, ez[20,:,:]' % repr(fields.ns)) ax1.set_xlabel('x') ax1.set_ylabel('z') ax2 = fig.add_subplot(2, 3, 5) ax2.imshow(fields.get('ex')[:,ny/2,:].T, vmin=-1.1, vmax=1.1) ax2.set_title('%s, ex[:,:,20]' % repr(fields.ns)) ax2.set_xlabel('x') ax2.set_ylabel('z') # x-axis nx, ny, nz = 2, 180, 160 fields = Fields(QueueTask(), nx, ny, nz) Core(fields) Pbc(fields, 'xyz') IncidentDirect(fields, 'ez', (0, 20, 0), (-1, 20, -1), tfunc) IncidentDirect(fields, 'ey', (0, 0, 20), (-1, -1, 20), tfunc) for tstep in xrange(1, tmax+1): fields.update_e() fields.update_h() fields.enqueue_barrier() ax1 = fig.add_subplot(2, 3, 3) ax1.imshow(fields.get('ez')[nx/2,:,:].T, vmin=-1.1, vmax=1.1) ax1.set_title('%s, ez[:,20,:]' % repr(fields.ns)) ax1.set_xlabel('y') ax1.set_ylabel('z') ax2 = fig.add_subplot(2, 3, 6) ax2.imshow(fields.get('ey')[nx/2,:,:].T, vmin=-1.1, vmax=1.1) ax2.set_title('%s, ey[:,:,20]' % repr(fields.ns)) ax2.set_xlabel('y') ax2.set_ylabel('z') #plt.savefig('./png/%.6d.png' % tstep) plt.show()
[ "kh.kim@kiaps.org" ]
kh.kim@kiaps.org
b04305f109a7f6222d39d666b9d65f277cae3196
ce972e94fcdf19d6809d94c2a73595233d1f741d
/catkin_ws/build/tf/cmake/tf-genmsg-context.py
552ade3ed0568d977725f0742c6cd2ff5e746bcd
[]
no_license
WilliamZipanHe/reward_shaping_ttr
cfa0e26579f31837c61af3e09621b4dad7eaaba2
df56cc0153147bb067bc3a0eee0e1e4e1044407f
refs/heads/master
2022-02-23T05:02:00.120626
2019-08-07T21:52:50
2019-08-07T21:52:50
null
0
0
null
null
null
null
UTF-8
Python
false
false
830
py
# generated from genmsg/cmake/pkg-genmsg.context.in messages_str = "/local-scratch/xlv/catkin_ws/src/geometry/tf/msg/tfMessage.msg" services_str = "/local-scratch/xlv/catkin_ws/src/geometry/tf/srv/FrameGraph.srv" pkg_name = "tf" dependencies_str = "geometry_msgs;sensor_msgs;std_msgs" langs = "gencpp;geneus;genlisp;gennodejs;genpy" dep_include_paths_str = "tf;/local-scratch/xlv/catkin_ws/src/geometry/tf/msg;geometry_msgs;/opt/ros/kinetic/share/geometry_msgs/cmake/../msg;sensor_msgs;/opt/ros/kinetic/share/sensor_msgs/cmake/../msg;std_msgs;/opt/ros/kinetic/share/std_msgs/cmake/../msg" PYTHON_EXECUTABLE = "/local-scratch/xlv/miniconda3/envs/py35_no_specific/bin/python3.5" package_has_static_sources = 'TRUE' == 'TRUE' genmsg_check_deps_script = "/opt/ros/kinetic/share/genmsg/cmake/../../../lib/genmsg/genmsg_check_deps.py"
[ "xlv@cs-mars-01.cmpt.sfu.ca" ]
xlv@cs-mars-01.cmpt.sfu.ca
c1f2cdad0c15be0461cfad4fc0ce8cad5a672d6f
eca90951b53822740812e40572e209728f71c261
/models.py
a109ee157f955568d9590b918a0010e3fee55a02
[]
no_license
tungvx/reporting
b7117879be773fccd90dbdb36a9e1220edc1d202
98f54c821aad761c0ab0ab83a8faad232ece1b41
refs/heads/master
2021-01-16T18:20:46.613907
2012-03-25T13:42:06
2012-03-25T13:42:06
3,491,504
0
0
null
null
null
null
UTF-8
Python
false
false
1,661
py
# -*- coding: utf-8 -*- import datetime from django.db import models from django import forms class Upload(models.Model): #Upload files table in databases filename = models.CharField(max_length=255) upload_time = models.DateTimeField('time uploaded') description = models.CharField(max_length=255) filestore = models.CharField(max_length=255) def __unicode__(self): return self.description class Spreadsheet_report(models.Model): # model to store the information about the spreadsheet used by user created_time = models.DateTimeField('time created') description = models.CharField(max_length=255) spreadsheet_link = models.CharField(max_length=255) output_link = models.CharField(max_length=255) title = models.CharField(max_length=255) def __unicode__(self): return self.description class upload_file_form(forms.Form): # Define a simple form for uploading excels file description = forms.CharField(max_length=255,required=True) file = forms.FileField(required=True,) def handle_uploaded_file(f,location,filename): #Save file upload content to uploaded folder fd = open('%s/%s' % (location, str(filename)), 'wb') #Create new file for write for chunk in f.chunks(): fd.write(chunk) #Write file data fd.close() #Close the file class spreadsheet_report_form(forms.Form): description = forms.CharField(max_length=255,required=True) spreadsheet_link = forms.CharField(max_length=255,required=False)
[ "toilatung90@gmail.com" ]
toilatung90@gmail.com
ba265bb9d96f3ceeec3b311c1c36ce36f9c18206
f50114692187a054bf2627695c6380d5ac79a20c
/q0028.py
26642f410f3a67d77ed2d7e0af65fad57a49ecc3
[]
no_license
petitepirate/interviewQuestions
c0cb8775932f90ff5c8e4ef80be468ef4155052f
209322c1f1ddbe8111dc2c5e9c35aaf787e0196a
refs/heads/master
2023-07-17T12:15:22.847440
2021-06-14T02:32:39
2021-06-14T02:32:39
286,884,022
1
0
null
null
null
null
UTF-8
Python
false
false
4,685
py
# This problem was asked by Palantir. # Write an algorithm to justify text. Given a sequence of words and an integer line length # k, return a list of strings which represents each line, fully justified. # More specifically, you should have as many words as possible in each line. There should # be at least one space between each word. Pad extra spaces when necessary so that each line # has exactly length k. Spaces should be distributed as equally as possible, with the extra # spaces, if any, distributed starting from the left. # If you can only fit one word on a line, then you should pad the right-hand side with spaces. # Each word is guaranteed not to be longer than k. # For example, given the list of words ["the", "quick", "brown", "fox", "jumps", "over", # "the", "lazy", "dog"] and k = 16, you should return the following: # ["the quick brown", # 1 extra space on the left # "fox jumps over", # 2 extra spaces distributed evenly # "the lazy dog"] # 4 extra spaces distributed evenly # ________________________________________________________________________________________ # Solution # It seems like the justification algorithm is independent from the groupings, so immediately # we should figure out two things: # How to group lines together so that it is as close to k as possible (without going over) # Given a grouping of lines, justifying the text by appropriately distributing spaces # To solve the first part, let's write a function group_lines that takes in all the words in # our input sequence as well as out target line length k, and return a list of list of words # that represents the lines that we will eventually justify. Our main strategy will be to # iterate over all the words, keep a list of words for the current line, and because we want # to fit as many words as possible per line, estimate the current line length, assuming only # one space between each word. Once we go over k, then save the word and start a new line with # it. So our function will look something like this: import math def min_line(words): return ' '.join(words) def group_lines(words, k): ''' Returns groupings of |words| whose total length, including 1 space in between, is less than |k|. ''' groups = [] current_sum = 0 current_line = [] for _, word in enumerate(words): # Check if adding the next word would push it over # the limit. If it does, then add |current_line| to # group. Also reset |current_line| properly. if len(min_line(current_line + [word])) > k: groups.append(current_line) current_line = [] current_line.append(word) # Add the last line to groups. groups.append(current_line) return groups # Then, we'll want to actually justify each line. We know for sure each line we feed # from group_lines is the maximum number of words we can pack into a line and no more. # What we can do is first figure out how many spaces we have available to distribute # between each word. Then from that, we can calculate how much base space we should # have between each word by dividing it by the number of words minus one. If there are # any leftover spaces to distribute, then we can keep track of that in a counter, and # as we rope in each new word we'll add the appropriate number of spaces. We can't add # more than one leftover space per word. def justify(words, length): ''' Precondition: |words| can fit in |length|. Justifies the words using the following algorithm: - Find the smallest spacing between each word (available_spaces / spaces) - Add a leftover space one-by-one until we run out ''' if len(words) == 1: word = words[0] num_spaces = length - len(word) spaces = ' ' * num_spaces return word + spaces spaces_to_distribute = length - sum(len(word) for word in words) number_of_spaces = len(words) - 1 smallest_space = math.floor(spaces_to_distribute / number_of_spaces) leftover_spaces = spaces_to_distribute - \ (number_of_spaces * smallest_space) justified_words = [] for word in words: justified_words.append(word) current_space = ' ' * smallest_space if leftover_spaces > 0: current_space += ' ' leftover_spaces -= 1 justified_words.append(current_space) return ''.join(justified_words).rstrip() # The final solution should just combine our two functions: def justify_text(words, k): return [justify(group, k) for group in group_lines(words, k)]
[ "msmeganmcmanus@gmail.com" ]
msmeganmcmanus@gmail.com
ede19815e880082d79936d9cd70e6f8e3e80d2d2
0ceb9ffeb2c087b8ae75c1a1179387fe36379f17
/test_users.py
31f33f5ba427527d4314a2b07fc4c576bc50f3d0
[]
no_license
chetat/eveno
47a3a99ebe4c5ea10252e1a21c45129e61b3e2ba
dbc138cde6e48039614cea52d3dc7bcad869a1dd
refs/heads/master
2022-12-09T13:51:46.923622
2021-03-09T10:21:25
2021-03-09T10:21:25
224,212,711
0
0
null
2022-09-16T18:20:10
2019-11-26T14:34:12
Python
UTF-8
Python
false
false
1,623
py
import os import json from app import create_app, sqlalchemy as db from sqlalchemy import create_engine, text from flask_sqlalchemy import SQLAlchemy from app.config import TestingConfig from models import initialize_db import unittest class UsersTestCase(unittest.TestCase): """This class represents the Event App test case""" def setUp(self): """Executed before each test. Define test variables and initialize app.""" self.app = create_app(TestingConfig) self.client = self.app.test_client self.user = { "email": "yeku@gmail.com", "firstname": "Yeku Wilfred", "lastname": "chetat", "phone": "671357962", "password": "weezybaby" } with self.app.app_context(): # create all tables db.create_all() initialize_db() def tearDown(self): with self.app.app_context(): db.session.remove() db.drop_all() def test_create_user(self): res = self.client().post("api/v1/users", json=self.user) self.assertTrue(res.status_code, 200) def test_get_users(self): res = self.client().get("api/v1/users") self.assertTrue(res.status_code, 200) def test_invalid_credentials(self): res = self.client().post("api/v1/auth", json={"email": "yekuwilfred@gmailcom", "password": "wybaby" }) self.assertEqual(res.status_code, 404) if __name__ == "__main__": unittest.main()
[ "yekuwilfred@gmail.com" ]
yekuwilfred@gmail.com
0cc046943de4f6e8e84509edc6055886e1602b1e
024a515e8741ecc88b4cc20c067a5ef7785375f2
/preproc_dataset.py
0241efc2f5bb8f421152bf5b2a64df8c226e9aa6
[ "MIT" ]
permissive
PCJohn/maml
e3752a842557210e955f68479e0a056766647d54
17362ed5f9d85863ead5774c3e08e163c260bd16
refs/heads/master
2020-03-28T11:37:42.014350
2018-10-01T17:08:12
2018-10-01T17:08:12
148,232,484
0
0
null
2018-09-10T23:35:04
2018-09-10T23:35:03
null
UTF-8
Python
false
false
1,756
py
import os import numpy as np from PIL import Image dataset = 'miniimagenet' train_count = 10 # train samples per class val_count = 590 # val samples per class mnist_dir = '/home/prithvi/dsets/MNIST/trainingSet/' omniglot_dir = '/home/prithvi/dsets/Omniglot/train' miniimagenet_dir = '/home/prithvi/dsets/miniimagenet/train' save_dir = './data' if dataset == 'mnist': data_dir = mnist_dir size = (28,28) elif dataset == 'omniglot': data_dir = omniglot_dir size = (28,28) elif dataset == 'miniimagenet': data_dir = miniimagenet_dir size = (84,84) if __name__ == '__main__': train_dir = os.path.join(save_dir,'metatrain') val_dir = os.path.join(save_dir,'metaval') if not (os.path.exists(train_dir)): os.system('mkdir '+train_dir) if not (os.path.exists(val_dir)): os.system('mkdir '+val_dir) for cls in os.listdir(data_dir): cls_dir = os.path.join(data_dir,cls) cls_train_dir = os.path.join(train_dir,cls) if not os.path.exists(cls_train_dir): os.system('mkdir '+cls_train_dir) cls_val_dir = os.path.join(val_dir,cls) if not os.path.exists(cls_val_dir): os.system('mkdir '+cls_val_dir) samples = map(lambda x:(Image.open(x).resize(size,resample=Image.LANCZOS), os.path.split(x)[-1]), [os.path.join(cls_dir,s) for s in os.listdir(cls_dir)]) np.random.shuffle(samples) train,val = samples[:train_count],samples[train_count:train_count+val_count] for s,fname in train: s.save(os.path.join(cls_train_dir,fname)) for s,fname in val: s.save(os.path.join(cls_val_dir,fname))
[ "prithvichakra@gmail.com" ]
prithvichakra@gmail.com
976fe04075eb0ae35fbceb4cf4f95288fd19f182
9a0ada115978e9600ad7f1eab65fcc8825f637cf
/work_in_progress/kezhi_paper/classification/intermedia_classification_WT_new.py
23736c730b2fe6d41790fd9dfaa294059b6a9fd6
[]
no_license
ver228/work-in-progress
c1971f8d72b9685f688a10e4c5a1b150fa0812da
ef5baecc324da4550f81edb0513d38f039ee3429
refs/heads/master
2018-12-16T22:18:55.457290
2018-09-14T09:27:49
2018-09-14T09:27:49
56,165,301
0
0
null
null
null
null
UTF-8
Python
false
false
1,223
py
# -*- coding: utf-8 -*- """ Created on Tue Jun 06 16:16:01 2017 @author: kezhili """ import numpy as np import tables from sklearn.ensemble import RandomForestClassifier from sklearn.model_selection import cross_val_score no_neuron = 1040 # step size in the mesh no_fea = 30 fea_no = 30 names = ["Nearest Neighbors", "RBF SVM", "Gaussian Process", "Decision Tree", "Random Forest"] if __name__ == '__main__': fdata = 'interm_LR_WTmemory_interm_result_30features_WT.hdf5' with tables.File(fdata, 'r') as fid: X = fid.get_node('/X3')[:] Y = fid.get_node('/Y3')[:] #because of the fix of a previous error in the dataset, #where 'LSJ1' and 'L5J1' are actually the same class Y[97:107] = Y[107] with open('result_30_features.txt') as f: lines = f.read().split('\n') X_ind = [x == 'True' for x in lines if x] Xp = X[:,X_ind] cross_validation_fold = 5 for n_estimators in [10, 100, 1000]: clf2 = RandomForestClassifier(n_estimators=n_estimators) c_val = cross_val_score(clf2, Xp, Y, cv = cross_validation_fold) print(np.mean(c_val), np.std(c_val))
[ "ver228@gmail.com" ]
ver228@gmail.com
694c9e252cceb994dcb65e0a0c42be5b6500d395
10e89eb922a5c122079a55234169e5b0e7af0819
/histore/cli/base.py
06f81c1b370a21633cbbc13230aeb3a0d81faa09
[ "BSD-3-Clause" ]
permissive
Sandy4321/histore
5753dd0008d1ae3400506181f22789aa9fdb43ba
d600052514a1c5f672137f76a6e1388184b17cd4
refs/heads/master
2023-04-15T07:57:49.762415
2021-04-28T11:12:31
2021-04-28T11:12:31
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,101
py
# This file is part of the History Store (histore). # # Copyright (C) 2018-2021 New York University. # # The History Store (histore) is released under the Revised BSD License. See # file LICENSE for full license details. """Command line interface to interact with a manager for archives on the local file system. """ import click import os import sys from histore.archive.manager.persist import PersistentArchiveManager from histore.cli.archive import create_archive, delete_archive, list_archives, rename_archive from histore.cli.snapshot import checkout_snapshot, commit_snapshot, list_snapshots import histore.config as config # -- Init the archive manager ------------------------------------------------- @click.command(name='init') @click.option( '-b', '--basedir', required=False, type=click.Path(file_okay=False, dir_okay=True), help='Base directory for archive files' ) @click.option( '-c', '--dbconnect', required=False, help='Connect URL for the database' ) def init_manager(basedir, dbconnect): """Initialize the archive manager directory.""" # Test if the base directory exists and is empty. basedir = basedir if basedir is not None else config.BASEDIR() if os.path.isdir(basedir): if os.listdir(basedir): click.echo('Not an empty directory {}.'.format(basedir)) sys.exit(-1) # Create instance of persistent archive manager to setup directories and # files. PersistentArchiveManager( basedir=basedir, dbconnect=dbconnect, create=True ) click.echo("Initialized in {}.".format(os.path.abspath(basedir))) # -- Create command group ----------------------------------------------------- @click.group() def cli(): # pragma: no cover """Command line interface for HISTORE archive manager.""" pass cli.add_command(init_manager) cli.add_command(checkout_snapshot) cli.add_command(commit_snapshot) cli.add_command(create_archive) cli.add_command(delete_archive) cli.add_command(list_archives) cli.add_command(list_snapshots) cli.add_command(rename_archive)
[ "heiko.muller@gmail.com" ]
heiko.muller@gmail.com
7b4029c8b2039d20fcfa67a38e0af84120540d04
21e35e3889cd0064474343a3b84aa289a01f8fac
/third_example_flask_app/classes.py
89d5863ab289c20518eb2bebec2bf9439ba3ca5f
[]
no_license
EricSchles/learn_python
b8fbc2b38647efb8e7176ac0d20708ffe1691007
8a5eb76aa333253a6c01f76d36dacad6bcf931ea
refs/heads/master
2021-01-13T16:45:53.606240
2017-01-31T21:10:53
2017-01-31T21:10:53
77,006,920
0
1
null
null
null
null
UTF-8
Python
false
false
368
py
class Integer: def __init__(self,number): self.number = number def add(self,other_integer): self.number += other_integer def subtract(self,other_integer): self.number -= other_integer def __str__(self): return repr(self.number) if __name__ == '__main__': number = Integer(5) number.add(7) print(number)
[ "ericschles@gmail.com" ]
ericschles@gmail.com
6c81ceed055c2d375b64725926e5208cb180263e
4e96f383d4703ad8ee58869ed91a0c8432c8a051
/Cura/Cura/plugins/VersionUpgrade/VersionUpgrade25to26/__init__.py
c74b3218b6ee2cb2b18f90149702826398f3ef7d
[ "GPL-3.0-only", "LGPL-3.0-only" ]
permissive
flight7788/3d-printing-with-moveo-1
b2dba26010c4fa31815bc1d2d0966161a8600081
7fcb9c6b5da9245d54ac917de8c2a7f5148e42b0
refs/heads/Feature_Marlin_with_AlanBoy
2022-08-30T18:36:44.785058
2020-05-30T07:52:58
2020-05-30T07:52:58
212,583,912
0
0
MIT
2020-05-16T07:39:47
2019-10-03T13:13:01
C
UTF-8
Python
false
false
2,027
py
# Copyright (c) 2018 Ultimaker B.V. # Cura is released under the terms of the LGPLv3 or higher. from typing import Any, Dict, TYPE_CHECKING from . import VersionUpgrade25to26 if TYPE_CHECKING: from UM.Application import Application upgrade = VersionUpgrade25to26.VersionUpgrade25to26() def getMetaData() -> Dict[str, Any]: return { "version_upgrade": { # From To Upgrade function ("preferences", 4000000): ("preferences", 4000001, upgrade.upgradePreferences), # NOTE: All the instance containers share the same general/version, so we have to update all of them # if any is updated. ("quality_changes", 2000000): ("quality_changes", 2000001, upgrade.upgradeInstanceContainer), ("user", 2000000): ("user", 2000001, upgrade.upgradeInstanceContainer), ("definition_changes", 2000000): ("definition_changes", 2000001, upgrade.upgradeInstanceContainer), ("machine_stack", 3000000): ("machine_stack", 3000001, upgrade.upgradeMachineStack), }, "sources": { "quality_changes": { "get_version": upgrade.getCfgVersion, "location": {"./quality"} }, "preferences": { "get_version": upgrade.getCfgVersion, "location": {"."} }, "user": { "get_version": upgrade.getCfgVersion, "location": {"./user"} }, "definition_changes": { "get_version": upgrade.getCfgVersion, "location": {"./machine_instances"} }, "machine_stack": { "get_version": upgrade.getCfgVersion, "location": {"./machine_instances"} } } } def register(app: "Application") -> Dict[str, Any]: return { "version_upgrade": upgrade }
[ "t106360212@ntut.org.tw" ]
t106360212@ntut.org.tw
6d9ced454145fc2b7691e9934b6e754a3ceb726d
18536f8145457a193b976eec44ee92535f588e54
/tests/functional/s3api/test_select_object_content.py
87750a30e5124bb75e76a8d9613cd5c8052055b8
[ "Apache-2.0" ]
permissive
jamsheedsaeed/awsapp
07e4ec6b9e07f679106db8e61f104ee9065e6af0
5498a3652b1d471a8695e14ca9739140e88a4b29
refs/heads/master
2023-01-01T23:44:16.181967
2020-06-19T06:56:48
2020-06-19T06:56:48
273,418,893
0
0
NOASSERTION
2022-12-26T20:16:39
2020-06-19T06:16:31
Python
UTF-8
Python
false
false
4,758
py
#!/usr/bin/env python # Copyright 2018 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"). You # may not use this file except in compliance with the License. A copy of # the License is located at # # http://aws.amazon.com/apache2.0/ # # or in the "license" file accompanying this file. This file is # distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF # ANY KIND, either express or implied. See the License for the specific # language governing permissions and limitations under the License. import os import tempfile import shutil from awscli.testutils import BaseAWSCommandParamsTest from awscli.testutils import BaseAWSHelpOutputTest class TestGetObject(BaseAWSCommandParamsTest): prefix = ['s3api', 'select-object-content'] def setUp(self): super(TestGetObject, self).setUp() self.parsed_response = {'Payload': self.create_fake_payload()} self._tempdir = tempfile.mkdtemp() def tearDown(self): super(TestGetObject, self).tearDown() shutil.rmtree(self._tempdir) def create_fake_payload(self): yield {'Records': {'Payload': b'a,b,c,d\n'}} # These next two events are ignored because they aren't # "Records". yield {'Progress': {'Details': {'BytesScanned': 1048576, 'BytesProcessed': 37748736}}} yield {'Records': {'Payload': b'e,f,g,h\n'}} yield {'Stats': {'Details': {'BytesProcessed': 62605400, 'BytesScanned': 1662276}}} yield {'End': {}} def test_can_stream_to_file(self): filename = os.path.join(self._tempdir, 'outfile') cmdline = self.prefix[::] cmdline.extend(['--bucket', 'mybucket']) cmdline.extend(['--key', 'mykey']) cmdline.extend(['--expression', 'SELECT * FROM S3Object']) cmdline.extend(['--expression-type', 'SQL']) cmdline.extend(['--request-progress', 'Enabled=True']) cmdline.extend(['--input-serialization', '{"CSV": {}, "CompressionType": "GZIP"}']) cmdline.extend(['--output-serialization', '{"CSV": {}}']) cmdline.extend([filename]) expected_params = { 'Bucket': 'mybucket', 'Key': u'mykey', 'Expression': 'SELECT * FROM S3Object', 'ExpressionType': 'SQL', 'InputSerialization': {'CSV': {}, 'CompressionType': 'GZIP'}, 'OutputSerialization': {'CSV': {}}, 'RequestProgress': {'Enabled': True}, } stdout = self.assert_params_for_cmd(cmdline, expected_params)[0] self.assertEqual(stdout, '') with open(filename, 'r') as f: contents = f.read() self.assertEqual(contents, ( 'a,b,c,d\n' 'e,f,g,h\n' )) def test_errors_are_propagated(self): self.http_response.status_code = 400 self.parsed_response = { 'Error': { 'Code': 'CastFailed', 'Message': 'Attempt to convert from one data type to another', } } cmdline = self.prefix + [ '--bucket', 'mybucket', '--key', 'mykey', '--expression', 'SELECT * FROM S3Object', '--expression-type', 'SQL', '--request-progress', 'Enabled=True', '--input-serialization', '{"CSV": {}, "CompressionType": "GZIP"}', '--output-serialization', '{"CSV": {}}', os.path.join(self._tempdir, 'outfile'), ] expected_params = { 'Bucket': 'mybucket', 'Key': u'mykey', 'Expression': 'SELECT * FROM S3Object', 'ExpressionType': 'SQL', 'InputSerialization': {'CSV': {}, 'CompressionType': 'GZIP'}, 'OutputSerialization': {'CSV': {}}, 'RequestProgress': {'Enabled': True}, } self.assert_params_for_cmd( cmd=cmdline, params=expected_params, expected_rc=254, stderr_contains=( 'An error occurred (CastFailed) when ' 'calling the SelectObjectContent operation'), ) class TestHelpOutput(BaseAWSHelpOutputTest): def test_output(self): self.driver.main(['s3api', 'select-object-content', 'help']) # We don't want to be super picky because the wording may change # We just want to verify the Output section was customized. self.assert_contains( 'Output\n======\n' 'This command generates no output' ) self.assert_not_contains('[outfile') self.assert_contains('outfile')
[ "jamsheedsaeed786@gmail.com" ]
jamsheedsaeed786@gmail.com
d9fe6f00109d7ea449564f348bea0fbcf2feca43
001184c168b93118f0429b11bab55fe108928b5d
/Week 3- Programming Assignments/Orientation2.py
533af81246e273eec244ce942ae5e3e9300f2085
[]
no_license
harrypotter0/ml-robot
caee1f9695427b8a83f64a41420f89b948fdb801
06e6672ba47fd73d4077ff5a6f48bdfafaea6597
refs/heads/master
2021-09-14T04:28:31.987037
2018-05-08T14:30:04
2018-05-08T14:30:04
115,803,231
1
0
null
null
null
null
UTF-8
Python
false
false
4,177
py
# In this exercise, write a program that will # run your previous code twice. # Please only modify the indicated area below! from math import * import random landmarks = [[20.0, 20.0], [80.0, 80.0], [20.0, 80.0], [80.0, 20.0]] world_size = 100.0 class robot: def __init__(self): self.x = random.random() * world_size self.y = random.random() * world_size self.orientation = random.random() * 2.0 * pi self.forward_noise = 0.0; self.turn_noise = 0.0; self.sense_noise = 0.0; def set(self, new_x, new_y, new_orientation): if new_x < 0 or new_x >= world_size: raise ValueError, 'X coordinate out of bound' if new_y < 0 or new_y >= world_size: raise ValueError, 'Y coordinate out of bound' if new_orientation < 0 or new_orientation >= 2 * pi: raise ValueError, 'Orientation must be in [0..2pi]' self.x = float(new_x) self.y = float(new_y) self.orientation = float(new_orientation) def set_noise(self, new_f_noise, new_t_noise, new_s_noise): # makes it possible to change the noise parameters # this is often useful in particle filters self.forward_noise = float(new_f_noise); self.turn_noise = float(new_t_noise); self.sense_noise = float(new_s_noise); def sense(self): Z = [] for i in range(len(landmarks)): dist = sqrt((self.x - landmarks[i][0]) ** 2 + (self.y - landmarks[i][1]) ** 2) dist += random.gauss(0.0, self.sense_noise) Z.append(dist) return Z def move(self, turn, forward): if forward < 0: raise ValueError, 'Robot cant move backwards' # turn, and add randomness to the turning command orientation = self.orientation + float(turn) + random.gauss(0.0, self.turn_noise) orientation %= 2 * pi # move, and add randomness to the motion command dist = float(forward) + random.gauss(0.0, self.forward_noise) x = self.x + (cos(orientation) * dist) y = self.y + (sin(orientation) * dist) x %= world_size # cyclic truncate y %= world_size # set particle res = robot() res.set(x, y, orientation) res.set_noise(self.forward_noise, self.turn_noise, self.sense_noise) return res def Gaussian(self, mu, sigma, x): # calculates the probability of x for 1-dim Gaussian with mean mu and var. sigma return exp(- ((mu - x) ** 2) / (sigma ** 2) / 2.0) / sqrt(2.0 * pi * (sigma ** 2)) def measurement_prob(self, measurement): # calculates how likely a measurement should be prob = 1.0; for i in range(len(landmarks)): dist = sqrt((self.x - landmarks[i][0]) ** 2 + (self.y - landmarks[i][1]) ** 2) prob *= self.Gaussian(dist, self.sense_noise, measurement[i]) return prob def __repr__(self): return '[x=%.6s y=%.6s orient=%.6s]' % (str(self.x), str(self.y), str(self.orientation)) #myrobot = robot() #myrobot.set_noise(5.0, 0.1, 5.0) #myrobot.set(30.0, 50.0, pi/2) #myrobot = myrobot.move(-pi/2, 15.0) #print myrobot.sense() #myrobot = myrobot.move(-pi/2, 10.0) #print myrobot.sense() #### DON'T MODIFY ANYTHING ABOVE HERE! ENTER/MODIFY CODE BELOW #### myrobot = robot() myrobot = myrobot.move(0.1, 5.0) Z = myrobot.sense() N = 1000 p = [] for i in range(N): x = robot() x.set_noise(0.05, 0.05, 5.0) p.append(x) ntimes = 28 for i in range(ntimes): p2 = [] for i in range(N): p2.append(p[i].move(0.1, 5.0)) p = p2 w = [] for i in range(N): w.append(p[i].measurement_prob(Z)) p3 = [] index = int(random.random() * N) beta = 0.0 mw = max(w) for i in range(N): beta += random.random() * 2.0 * mw while beta > w[index]: beta -= w[index] index = (index + 1) % N p3.append(p[index]) p = p3 print p #Leave this print statement for grading purposes!
[ "9654263057akashkandpal@gmail.com" ]
9654263057akashkandpal@gmail.com
d91241e28a7781f4e1b8d3e0aaae4e0162f0622f
a3eccc652f83815318bdb033a33573c5b1e073e9
/nac/crm/views/add_lead.py
7d6cf4cdf3c36cb198817a0d24a038fb42afd38d
[]
no_license
jsvelu/coms-dev--old
8139fa511e2985b4d71550f1c59402069d09edf3
de300ad6ef947d29380972a6efe809f4ef05d7e1
refs/heads/main
2023-07-17T20:44:36.101738
2021-09-04T21:56:38
2021-09-04T21:56:38
403,158,834
0
0
null
null
null
null
UTF-8
Python
false
false
1,615
py
from authtools.views import resolve_url_lazy from django.urls import reverse_lazy from django.views.generic.edit import FormView from rules.contrib.views import PermissionRequiredMixin from crm.forms.lead import LeadForm from customers.models import Customer from dealerships.models import Dealership class AddLeadView(PermissionRequiredMixin, FormView): template_name = 'crm/add_lead.html' form_class = LeadForm success_url = reverse_lazy('crm:lead_listing') raise_exception = True permission_required = 'customers.add_customer' def get_allowed_dealerships(self): dealer_choices = None if self.request.user.has_perm('customers.manage_self_and_dealership_leads_only'): dealer_choices = Dealership.objects.filter(dealershipuser=self.request.user) if self.request.user.has_perm('crm.manage_all_leads'): dealer_choices = Dealership.objects.all() return dealer_choices def get_form_kwargs(self): kwargs = super(AddLeadView, self).get_form_kwargs() dealer_choices = [(dealership.id, dealership.name) for dealership in self.get_allowed_dealerships()] kwargs.update({'dealership_choices': dealer_choices}) return kwargs def get_context_data(self, **kwargs): context = super(AddLeadView, self).get_context_data(**kwargs) context['sub_heading'] = 'Add Lead' return context def form_valid(self, form): lead = form.save(commit=False) lead.lead_type = Customer.LEAD_TYPE_LEAD lead.save() return super(AddLeadView, self).form_valid(form)
[ "velu@qrsolutions.in" ]
velu@qrsolutions.in
77388ef09898130d90f9cc56214733cdd9160d06
4b60c34ba37e7c0611257e7934791fb43d01e254
/src/Lib/pysparseSuperLU.py
2524ee2635966f53d120c079893e98ac0bbfb303
[]
no_license
regmi/pysparse
0913ff69b5d07b58c20deb5b6f44caeaa8498d64
ebc2ad045382c69e6bb41217c9431e51736ac4a0
refs/heads/master
2021-01-01T16:20:10.075871
2010-04-09T22:56:51
2010-04-09T22:56:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
5,502
py
""" A framework for solving sparse linear systems of equations using an LU factorization, by means of the supernodal sparse LU factorization package SuperLU ([DEGLL99]_, [DGL99]_, [LD03]_). This package is appropriate for factorizing sparse square unsymmetric or rectangular matrices. See [SLU]_ for more information. **References:** .. [DEGLL99] J. W. Demmel, S. C. Eisenstat, J. R. Gilbert, X. S. Li and J. W. H. Liu, *A supernodal approach to sparse partial pivoting*, SIAM Journal on Matrix Analysis and Applications **20**\ (3), pp. 720-755, 1999. .. [DGL99] J. W. Demmel, J. R. Gilbert and X. S. Li, *An Asynchronous Parallel Supernodal Algorithm for Sparse Gaussian Elimination*, SIAM Journal on Matrix Analysis and Applications **20**\ (4), pp. 915-952, 1999. .. [LD03] X. S. Li and J. W. Demmel, *SuperLU_DIST: A Scalable Distributed-Memory Sparse Direct Solver for Unsymmetric Linear Systems*, ACM Transactions on Mathematical Software **29**\ (2), pp. 110-140, 2003. .. [SLU] http://crd.lbl.gov/~xiaoye/SuperLU """ # To look into: # - allow other data types __docformat__ = 'restructuredtext' import pysparseMatrix as psm import numpy import resource from directSolver import PysparseDirectSolver from pysparse import superlu def cputime(): return resource.getrusage(resource.RUSAGE_SELF)[0] class PysparseSuperLUSolver( PysparseDirectSolver ): """ `PysparseSuperLUSolver` is a wrapper class around the SuperLu library for the factorization of full-rank n-by-m matrices. Only matrices with real coefficients are currently supported. :parameters: :A: The matrix to be factorized, supplied as a PysparseMatrix instance. :keywords: :symmetric: a boolean indicating that the user wishes to use symmetric mode. In symmetric mode, ``permc_spec=2`` must be chosen and ``diag_pivot_thresh`` must be small, e.g., 0.0 or 0.1. Since the value of ``diag_pivot_thresh`` is up to the user, setting ``symmetric`` to ``True`` does *not* automatically set ``permc_spec`` and ``diag_pivot_thresh`` to appropriate values. :diag_pivot_thresh: a float value between 0 and 1 representing the threshold for partial pivoting (0 = no pivoting, 1 = always perform partial pivoting). Default: 1.0. :drop_tol: the value of a drop tolerance, between 0 and 1, if an incomplete factorization is desired (0 = exact factorization). This keyword does not exist if using SuperLU version 2.0 and below. In more recent version of SuperLU, the keyword is accepted but has no effect. Default: 0.0 :relax: an integer controling the degree of relaxing supernodes. Default: 1. :panel_size: an integer specifying the maximum number of columns to form a panel. Default: 10. :permc_spec: an integer specifying the ordering strategy used during the factorization. 0. natural ordering, 1. MMD applied to the structure of :math:`\mathbf{A}^T \mathbf{A}` 2. MMD applied to the structure of :math:`\mathbf{A}^T + \mathbf{A}` 3. COLAMD. Default: 2. .. attribute:: LU A :class:`superlu_context` object encapsulating the factorization. .. attribute:: sol The solution of the linear system after a call to :meth:`solve`. .. attribute:: factorizationTime The CPU time to perform the factorization. .. attribute:: solutionTime The CPU time to perform the forward and backward sweeps. .. attribute:: lunz The number of nonzero elements in the factors L and U together after a call to :meth:`fetch_lunz`. """ def __init__(self, A, **kwargs): PysparseDirectSolver.__init__(self, A, **kwargs) self.type = numpy.float self.nrow, self.ncol = A.getShape() t = cputime() self.LU = superlu.factorize(A.matrix.to_csr(), **kwargs) self.factorizationTime = cputime() - t self.solutionTime = 0.0 self.sol = None self.L = self.U = None return def solve(self, rhs, transpose = False): """ Solve the linear system ``A x = rhs``, where ``A`` is the input matrix and ``rhs`` is a Numpy vector of appropriate dimension. The result is placed in the :attr:`sol` member of the class instance. If the optional argument ``transpose`` is ``True``, the transpose system ``A^T x = rhs`` is solved. """ if self.sol is None: self.sol = numpy.empty(self.ncol, self.type) transp = 'N' if transpose: transp = 'T' t = cputime() self.LU.solve(rhs, self.sol, transp) self.solutionTime = cputime() - t return def fetch_lunz(self): """ Retrieve the number of nonzeros in the factors L and U together. The result is stored in the member :attr:`lunz` of the class instance. """ self.lunz = self.LU.nnz def fetch_factors(self): """ Not yet available. """ raise NotImplementedError
[ "regmisk@gmail.com" ]
regmisk@gmail.com
5e3f127541774244db776a2ad4ca2f70cd1d14b9
dfdbc9118742bc09d7c7fe6fe42f53b7d1d7977a
/spacegame_ii/parralax.py
3ef67de9c8ce5c05576be44f185131d62e7cf4e6
[]
no_license
602p/spacegame
d1c3a34233ed7c7128d5cbe4c470c168a84700ac
1350beeb6df2b65a0c041f512fa944cbae4dba2b
refs/heads/master
2021-01-10T19:28:19.220218
2015-04-27T20:37:14
2015-04-27T20:37:14
27,309,904
0
0
null
2015-04-20T15:12:11
2014-11-29T17:39:59
Python
UTF-8
Python
false
false
1,454
py
from __future__ import division import pygame, random import logging module_logger=logging.getLogger("sg.parralax") debug, info, warning, error, critical = module_logger.debug, module_logger.info, module_logger.warning, module_logger.error, module_logger.critical class StarfieldLayer: def __init__(self, density, color, size, speed): self.density=density self.speed=speed self.particle_surface=pygame.Surface((size, size)) self.particle_surface.fill(color) def bind(self, (xsize, ysize)): self.particles=[] self.size=(xsize, ysize) def render(self, surface, (xpos, ypos)): i=0 state=random.getstate() random.seed(413) while i!=self.density: # pos=(random.uniform(0, self.size[0]), random.uniform(0, self.size[1])) surface.blit(self.particle_surface, ( int((((xpos+pos[0])/self.speed)%self.size[0])), int((((ypos+pos[1])/self.speed)%self.size[1])) ) ) i+=1 random.setstate(state) class ParralaxStarfieldScroller: def __init__(self, size, layers): self.layers=layers self.pos=[0,0] self.bindall(size) debug("Initilized ParralaxStarfieldScroller with "+str(len(layers))+" layers at []:"+str(size)) def bindall(self, size): for layer in self.layers: layer.bind(size) def render(self, surface): for layer in self.layers: layer.render(surface, self.pos) def move(self, x, y): self.pos[0]+=x self.pos[1]+=y def move_to(self, x, y): self.pos[0]=x self.pos[1]=y
[ "louis@goessling.com" ]
louis@goessling.com
fe3657f499c9f10b71a26ef58326f2f95d5634cb
6c4486ab599fd5dea9006e41cdb89db54b47b77c
/tests/products/NGP_OBLIC_Create_Ballpark.py
92f3d21ced7693e315f188073ea8a95c59c80077
[]
no_license
kenito2050/Python-Page-Object-Framework-Example
28ba61cdc1498374be4fc088a1348e0acb754dc2
2a3a3e6c74dc7ec7c9acce41030e9487925b9b0c
refs/heads/master
2020-04-02T15:52:24.286208
2018-10-25T01:46:24
2018-10-25T01:46:24
154,587,117
0
0
null
null
null
null
UTF-8
Python
false
false
5,830
py
from xml.etree import ElementTree as ET import xlrd import time from pages.producer_center.ballpark.ballpark_Indication import BallPark_Indication from pages.producer_center.ballpark.ballpark_PAF import BallPark_PAF from pages.producer_center.ballpark.ballpark_download_send import BallPark_Download_Send from pages.producer_center.products_programs_page import ProductsAndPrograms from pages.service_center.agents_page import AgentsPage from pages.service_center.login_page import LoginPage from pages.service_center.navigation_bar import NavigationBar from utilities.Environments.Environments import Environments from utilities.state_capitals.state_capitals import StateCapitals from utilities.zip_codes_state_capitals.zip_codes import ZipCodes from utilities.Faker.Data_Generator import Data_Generator from utilities.Date_Time_Generator.Date_Time_Generator import Date_Time_Generator from config_globals import * class TestCreateQuote(): def test_login_search_for_agent_create_quote(self, browser, env): Product = "NGP_OBLIC" driver = browser ## Directory Locations tests_directory = ROOT_DIR / 'tests' framework_directory = ROOT_DIR config_file_directory = CONFIG_PATH test_case_directory = framework_directory / 'utilities' / 'Excel_Sheets' / 'Products' test_results_directory = framework_directory / 'utilities' / 'Excel_Sheets' / 'Test_Results' global test_summary global test_scenario global effective_date global contract_class global agent global state global revenue global total_num_records global _OLD_scenario global limit global deductible # Open Test Scenario Workbook; Instantiate worksheet object # 0 - First Worksheet # 1 - Second Worksheet...etc wb = xlrd.open_workbook(str(test_case_directory / Product) + '.xlsx') sh = wb.sheet_by_index(3) ## Begin For Loop to iterate through Test Scenarios i = 1 rows = sh.nrows empty_cell = False for i in range(1, sh.nrows): cell_val = sh.cell(i, 0).value if cell_val == '': # If Cell Value is empty, set empty_cell to True empty_cell = True else: # If Cell Value is NOT empty, set empty_cell to False empty_cell = False # Check to see if cell is NOT empty # If cell is not empty, read in the values if empty_cell == False: test_summary = sh.cell_value(i, 0) test_scenario = str(round(sh.cell_value(i, 1))) effective_date = sh.cell_value(i, 2) contract_class = sh.cell_value(i, 3) agent = sh.cell_value(i, 4) state = sh.cell_value(i, 5) revenue = str(round(sh.cell_value(i, 6))) total_num_records = (sh.cell_value(i, 7)) _OLD_scenario = sh.cell_value(i, 8) limit = sh.cell_value(i, 9) deductible = sh.cell_value(i, 10) # Else, the cell is empty # End the Loop else: break # Create Instance of Data Generator dg = Data_Generator() # Create Company Name Value company_name_string = dg.create_full_company_name() # Create Street Address Value address_value = dg.create_street_address() city = StateCapitals.return_state_capital(state) postal_code = ZipCodes.return_zip_codes(state) # Create Instance of Date Time Generator dtg = Date_Time_Generator() # Create Today's Date date_today = dtg.return_date_today() # Access XML to retrieve login credentials tree = ET.parse(str(config_file_directory / 'resources.xml')) login_credentials = tree.getroot() username = (login_credentials[1][0].text) password = (login_credentials[1][1].text) ## Test Environment ## Select Appropriate URL based on the Environment Value (env) baseURL = Environments.return_environments(env) # Maximize Window; Launch URL driver.get(baseURL) driver.implicitly_wait(3) # Call Login methods from Pages.home.login_page.py lp = LoginPage(driver) lp.login(username, password) lp.click_login_button() nb = NavigationBar(driver) nb.click_agents() ap = AgentsPage(driver) ap.search_for_agent(agent) ap.click_submit_new_application_as_agent() pp = ProductsAndPrograms(driver) pp.click_ballpark() bp_PAF = BallPark_PAF(driver) bp_PAF.switch_windows() bp_PAF.start_ballpark_enter_faker_company_name_valid_zip(company_name_string, postal_code) bp_PAF.select_contract_class(contract_class) bp_PAF.click_ballpark_button() bp_PAF.select_NGP_OBLIC() time.sleep(3) # Enter Ad Hoc Effective Date # bp_PAF.enter_effective_date(ad_hoc_effectiveDate) # Enter Today's Date as Effective Date bp_PAF.enter_current_date(date_today) time.sleep(3) bp_PAF.enter_revenue(revenue) bp_PAF.click_ballpark_button() bp_Indication = BallPark_Indication(driver) bp_Indication.click_Download_Send_Indication() bp_Download_Send = BallPark_Download_Send(driver) bp_Download_Send.input_email() bp_Download_Send.click_send_email() # Close Ballpark Window driver.close() # Switch to First Window (Service Center) driver.switch_to.window(driver.window_handles[0]) # Wait driver.implicitly_wait(3) # Close Browser driver.quit()
[ "kvillarruel@nasinsurance.com" ]
kvillarruel@nasinsurance.com
ce8bc8e8c5475a8d2ed27e43305f5fd02ca3d509
da481ac79daaa68df0219e92b0d80a1a7aed1bf6
/python/example/run_all_tests.py
747477c1ca24629cf95203798b913aa2369a40a8
[]
no_license
PeterZhouSZ/diff_pd
5ebad8f8d361a3ac4b8c7fb25c923b85e01c8a31
45bf74bc1b601d0aba7cc5becc3d2582136b5c0b
refs/heads/master
2023-07-06T10:56:59.660391
2021-08-13T04:21:07
2021-08-13T04:21:07
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,144
py
import sys sys.path.append('../') from importlib import import_module from py_diff_pd.common.common import print_ok, print_error if __name__ == '__main__': # If you want to add a new test, simply add its name here --- you can find their names from README.md. tests = [ # Utilities. 'render_quad_mesh', # Numerical check. 'actuation_2d', 'actuation_3d', 'collision_2d', 'deformable_backward_2d', 'deformable_backward_3d', 'deformable_quasi_static_3d', 'pd_energy_2d', 'pd_energy_3d', 'pd_forward', 'state_force_2d', 'state_force_3d' ] failure_cnt = 0 for name in tests: test_func_name = 'test_{}'.format(name) module_name = name test_func = getattr(import_module(module_name), test_func_name) if test_func(verbose=False): print_ok('[{}] PASSED.'.format(name)) else: print_error('[{}] FAILED.'.format(name)) failure_cnt += 1 print('{}/{} tests failed.'.format(failure_cnt, len(tests))) if failure_cnt > 0: sys.exit(-1)
[ "taodu@csail.mit.edu" ]
taodu@csail.mit.edu
2958eecfff5d168b9defe4a24e536c251c58de46
db331fb24e5b95131413c8d5cebc880674dd30f7
/foundation/migrations/0009_title_img2.py
d49553924ebcef455145509d19762fb843f2a4e0
[]
no_license
kmvit/personalsite
78fb18f428e3c95219d9145cb22bdefdf8896bf9
5da20e853055affdad4e6d21f36bae6cfe7507b7
refs/heads/master
2021-06-18T13:34:03.800622
2017-04-07T12:40:24
2017-04-07T12:40:24
27,366,313
0
0
null
null
null
null
UTF-8
Python
false
false
464
py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('foundation', '0008_auto_20141201_2156'), ] operations = [ migrations.AddField( model_name='title', name='img2', field=models.ImageField(null=True, upload_to=b'media', blank=True), preserve_default=True, ), ]
[ "kmv-it@yandex.ru" ]
kmv-it@yandex.ru
79b729366e1e6dd4733ec4c9c4f4199ddbfb9d6a
7b91755b1c777248050f3cadf23ed34d1f10adef
/Section3/14.py
86af53d07f05b5c224b8112a78a39b964bf71c05
[]
no_license
JAntonioMarin/PythonBootcamp
ef2976be0204df44e0c56a521628c73a5d274008
6e4af15b725913d3fda60599792e17d7d43d61d2
refs/heads/master
2021-01-06T16:52:57.537678
2020-03-26T17:36:50
2020-03-26T17:36:50
241,406,222
0
0
null
null
null
null
UTF-8
Python
false
false
236
py
print('hello') print("world") print('this is also a string') print(" I'm going on a run ") print('hello \nworld') print('hello \tworld') print(len('hello')) print(len('I am')) # Coding Exercise 2: Quic Print Check print("Hello World")
[ "avalanch.psp@gmail.com" ]
avalanch.psp@gmail.com
9591b223d52310912c74667e3148d1398f10c830
124fe233f9cc86898756b3c0fc2988c69001f670
/tests/test_stat.py
212f9eec83cf924d08398be7ee3032045c77594c
[ "MIT" ]
permissive
jon-rd/jc
0c4e043ccc40fdf7a0771d48cec4d86321fc4588
1e18dd30a824b0463f0cad86e0da7094c47d34f9
refs/heads/master
2023-04-16T01:39:16.961164
2021-03-05T19:50:37
2021-03-05T19:50:37
351,145,638
0
0
MIT
2021-04-14T14:13:28
2021-03-24T16:14:42
null
UTF-8
Python
false
false
2,659
py
import os import json import unittest import jc.parsers.stat THIS_DIR = os.path.dirname(os.path.abspath(__file__)) class MyTests(unittest.TestCase): def setUp(self): # input with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/centos-7.7/stat.out'), 'r', encoding='utf-8') as f: self.centos_7_7_stat = f.read() with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/ubuntu-18.04/stat.out'), 'r', encoding='utf-8') as f: self.ubuntu_18_4_stat = f.read() with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/osx-10.14.6/stat.out'), 'r', encoding='utf-8') as f: self.osx_10_14_6_stat = f.read() with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/freebsd12/stat.out'), 'r', encoding='utf-8') as f: self.freebsd12_stat = f.read() # output with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/centos-7.7/stat.json'), 'r', encoding='utf-8') as f: self.centos_7_7_stat_json = json.loads(f.read()) with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/ubuntu-18.04/stat.json'), 'r', encoding='utf-8') as f: self.ubuntu_18_4_stat_json = json.loads(f.read()) with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/osx-10.14.6/stat.json'), 'r', encoding='utf-8') as f: self.osx_10_14_6_stat_json = json.loads(f.read()) with open(os.path.join(THIS_DIR, os.pardir, 'tests/fixtures/freebsd12/stat.json'), 'r', encoding='utf-8') as f: self.freebsd12_stat_json = json.loads(f.read()) def test_stat_nodata(self): """ Test 'stat' with no data """ self.assertEqual(jc.parsers.stat.parse('', quiet=True), []) def test_stat_centos_7_7(self): """ Test 'stat /bin/*' on Centos 7.7 """ self.assertEqual(jc.parsers.stat.parse(self.centos_7_7_stat, quiet=True), self.centos_7_7_stat_json) def test_stat_ubuntu_18_4(self): """ Test 'stat /bin/*' on Ubuntu 18.4 """ self.assertEqual(jc.parsers.stat.parse(self.ubuntu_18_4_stat, quiet=True), self.ubuntu_18_4_stat_json) def test_stat_osx_10_14_6(self): """ Test 'stat /foo/*' on OSX 10.14.6 """ self.assertEqual(jc.parsers.stat.parse(self.osx_10_14_6_stat, quiet=True), self.osx_10_14_6_stat_json) def test_stat_freebsd12(self): """ Test 'stat /foo/*' on FreeBSD12 """ self.assertEqual(jc.parsers.stat.parse(self.freebsd12_stat, quiet=True), self.freebsd12_stat_json) if __name__ == '__main__': unittest.main()
[ "kellyjonbrazil@gmail.com" ]
kellyjonbrazil@gmail.com
3e8474ef201455aa5739a31fe47a4749d1b82850
e6f4e3afd16a7ee5c7a8fb61f7ed697ce88ef4c4
/Pro2/abcpp/abcpp/treeplot.py
c8a4ebaadf40cfd0ca32ddbb58e1a29142193621
[]
no_license
xl0418/Code
01b58d05f7fae1a5fcfec15894ce0ed8c833fd1a
75235b913730714d538d6d822a99297da54d3841
refs/heads/master
2021-06-03T21:10:31.578731
2020-11-17T07:50:48
2020-11-17T07:50:48
136,896,128
0
0
null
null
null
null
UTF-8
Python
false
false
5,549
py
import sys, os import platform if platform.system() == 'Windows': sys.path.append('C:/Liang/abcpp_master8/abcpp') elif platform.system() == 'Darwin': sys.path.append('/Users/dudupig/Documents/GitHub/Code/Pro2/Python_p2') from dvtraitsim_py import DVSim from dvtraitsim_shared import DVTreeData, DVParam import numpy as np import matplotlib.pyplot as plt theta = 0 # optimum of natural selection r = 1 # growth rate Vmax = 1 scalar = 10000 K = 10e8 nu = 1 / (100 * K) timegap = 100 # let's try to find a true simulation: # trait evolution plot for no_tree in range(4, 23): gamma_vec = np.array([0, 0.001, 0.01, 0.1, 0.5, 1]) a_vec = gamma_vec row_gamma = len(gamma_vec) count = 0 tree = 'tree' + '%d' % no_tree example = 'example' + '%d' % no_tree if platform.system() == 'Windows': dir_path = 'c:/Liang/Googlebox/Research/Project2' files = dir_path + '/treesim_newexp/' + example + '/' td = DVTreeData(path=files, scalar=scalar) elif platform.system() == 'Darwin': file = '/Users/dudupig/Documents/GitHub/Code/Pro2/abcpp/tree_data/' + example + '/' f1, axes1 = plt.subplots(row_gamma, row_gamma, figsize=(9, 9), sharey=True, sharex=True) # f2, axes2 = plt.subplots(row_gamma, row_gamma, figsize=(9, 9), sharey=True, sharex=True) # f3, axes3 = plt.subplots(row_gamma, row_gamma, figsize=(9, 9), sharey=True, sharex=True) # label_a = (['$\\alpha$=0', '$\\alpha$=.001', '$\\alpha$=.01', '$\\alpha$=.1', '$\\alpha$=.5', '$\\alpha$=1']) label_gamma = (['$\gamma$=0', '$\gamma$=.001', '$\gamma$=.01', '$\gamma$=.1', '$\gamma$=.5', '$\gamma$=1']) xticks = (0, td.evo_time * scalar / timegap) xlabels = ['0', '150K'] for index_g in range(len(gamma_vec)): gamma1 = gamma_vec[index_g] for index_a in range(len(a_vec)): a = a_vec[index_a] print(count) if index_a >= index_g: for replicate in range(100): obs_param = DVParam(gamma=gamma1, a=a, K=K, nu=nu, r=r, theta=theta, Vmax=1, inittrait=0, initpop=500, initpop_sigma=10.0, break_on_mu=False) simresult = DVSim(td, obs_param) if simresult['sim_time'] == td.sim_evo_time: pic = 0 break else: pic = 1 # if pic==0: evo_time, total_species = simresult['N'].shape evo_time = evo_time - 1 trait_RI_dr = simresult['Z'] population_RI_dr = simresult['N'] population_RI_dr = population_RI_dr.astype(float) population_RI_dr[population_RI_dr == 0] = np.nan V_dr = simresult['V'] num_lines = total_species x = np.arange(evo_time / timegap + 1) labels = [] for i in range(1, num_lines + 1): axes1[index_g, index_a].plot(x, trait_RI_dr[::timegap, i - 1]) axes2[index_g, index_a].plot(x, population_RI_dr[::timegap, i - 1]) axes3[index_g, index_a].plot(x, V_dr[::timegap, i - 1]) axes1[index_g, index_a].set_xticks(xticks) axes1[index_g, index_a].set_xticklabels(xlabels, minor=False) axes2[index_g, index_a].set_xticks(xticks) axes2[index_g, index_a].set_xticklabels(xlabels, minor=False) axes3[index_g, index_a].set_xticks(xticks) axes3[index_g, index_a].set_xticklabels(xlabels, minor=False) if count in range(0, row_gamma): axes1[index_g, index_a].title.set_text(label_a[count]) axes2[index_g, index_a].title.set_text(label_a[count]) axes3[index_g, index_a].title.set_text(label_a[count]) if count in ([5, 11, 17, 23, 29, 35]): axes1[index_g, index_a].set_ylabel(label_gamma[int(count / row_gamma)]) axes1[index_g, index_a].yaxis.set_label_position("right") axes2[index_g, index_a].set_ylabel(label_gamma[int(count / row_gamma)]) axes2[index_g, index_a].yaxis.set_label_position("right") axes3[index_g, index_a].set_ylabel(label_gamma[int(count / row_gamma)]) axes3[index_g, index_a].yaxis.set_label_position("right") else: axes1[index_g, index_a].plot([]) axes2[index_g, index_a].plot([]) axes3[index_g, index_a].plot([]) axes1[index_g, index_a].axis('off') axes2[index_g, index_a].axis('off') axes3[index_g, index_a].axis('off') count += 1 dir_fig = 'C:/Liang/Googlebox/Research/Project2/smc_treeuppertri/' + tree f1.text(0.84, 0.04, 'Generation', ha='center', fontsize=15) f1.text(0.04, 0.84, 'Trait mean', va='center', rotation='vertical', fontsize=15) f2.text(0.84, 0.04, 'Generation', ha='center', fontsize=15) f2.text(0.04, 0.84, 'Population size', va='center', rotation='vertical', fontsize=15) f3.text(0.84, 0.04, 'Generation', ha='center', fontsize=15) f3.text(0.04, 0.84, 'Trait variance', va='center', rotation='vertical', fontsize=15) f1.savefig(dir_fig + 'TP.png') plt.close(f1) f2.savefig(dir_fig + 'NP.png') plt.close(f2) f3.savefig(dir_fig + 'VP.png') plt.close(f3) plt.close('all')
[ "xl0418@gmail.com" ]
xl0418@gmail.com
a823d40e45f948799c58daf72f5a9314e5da1b4c
25e481ef7fba79285f4c8a7fa2e81c8b2b7f9cce
/saleor/core/permissions.py
093869518cad2dc41e345780942a66464273b61a
[ "BSD-2-Clause" ]
permissive
arslanahmd/Ghar-Tameer
59e60def48a14f9452dfefe2edf30e362878191d
72401b2fc0079e6d52e844afd8fcf57122ad319f
refs/heads/master
2023-01-31T04:08:26.288332
2018-06-07T18:02:01
2018-06-07T18:02:01
136,231,127
0
0
NOASSERTION
2023-01-11T22:21:42
2018-06-05T20:28:11
Python
UTF-8
Python
false
false
957
py
from django.contrib.auth.models import Permission MODELS_PERMISSIONS = [ 'order.view_order', 'order.edit_order', 'product.view_category', 'product.edit_category', 'product.view_product', 'product.edit_product', 'product.view_properties', 'product.edit_properties', 'product.view_stock_location', 'product.edit_stock_location', 'sale.view_sale', 'sale.edit_sale', 'shipping.view_shipping', 'shipping.edit_shipping', 'site.edit_settings', 'site.view_settings', 'user.view_user', 'user.edit_user', 'user.view_group', 'user.edit_group', 'user.view_staff', 'user.edit_staff', 'user.impersonate_user', 'voucher.view_voucher', 'voucher.edit_voucher', ] def get_permissions(): codenames = [permission.split('.')[1] for permission in MODELS_PERMISSIONS] return Permission.objects.filter(codename__in=codenames)\ .prefetch_related('content_type')
[ "arslanahmad085@gmail.com" ]
arslanahmad085@gmail.com
9a9f2f15e46689e698a6cf1af6e148036e507196
0444e53f4908454e2e8ab9f70877ec76be9f872c
/reportng/migrations/0025_auto_20161204_2336.py
76552396407602970196277f9a8d5e618d06029d
[ "MIT" ]
permissive
dedayoa/keepintouch
c946612d6b69a20a92617354fef2ce6407382be5
2551fc21bb1f6055ab0fc2ec040c6ba874c2838f
refs/heads/master
2020-03-22T12:53:44.488570
2018-08-17T13:00:51
2018-08-17T13:00:51
140,069,262
1
1
null
null
null
null
UTF-8
Python
false
false
487
py
# -*- coding: utf-8 -*- # Generated by Django 1.9.10 on 2016-12-04 22:36 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('reportng', '0024_auto_20161204_2123'), ] operations = [ migrations.AlterField( model_name='emaildeliveryreport', name='sent_at', field=models.DateTimeField(null=True, verbose_name='Sent'), ), ]
[ "dedayoa@gmail.com" ]
dedayoa@gmail.com
f6200c41fbfc26030afb9f8196b7c6948fbc18be
ecf0d106831b9e08578845674a457a166b6e0a14
/OOP/inheritance_EXERCISE/restaurant/project/food/dessert.py
6f120937221db42a44dd44819bf082a0f5df016c
[]
no_license
ivo-bass/SoftUni-Solutions
015dad72cff917bb74caeeed5e23b4c5fdeeca75
75612d4bdb6f41b749e88f8d9c512d0e00712011
refs/heads/master
2023-05-09T23:21:40.922503
2021-05-27T19:42:03
2021-05-27T19:42:03
311,329,921
8
5
null
null
null
null
UTF-8
Python
false
false
288
py
from project.food.food import Food class Dessert(Food): def __init__(self, name: str, price: float, grams: float, calories: float): super().__init__(name, price, grams) self.__calories = calories @property def calories(self): return self.__calories
[ "ivailo.ignatoff@gmail.com" ]
ivailo.ignatoff@gmail.com
b0dd547c60357b7f4540d9f1064f30da3da49bfb
3b89c0a97ac6b58b6923a213bc8471e11ad4fe69
/python/CodingExercises/Median.py
9fca643773716391f6ec77c8b200529cd534e8b8
[]
no_license
ksayee/programming_assignments
b187adca502ecf7ff7b51dc849d5d79ceb90d4a6
13bc1c44e1eef17fc36724f20b060c3339c280ea
refs/heads/master
2021-06-30T07:19:34.192277
2021-06-23T05:11:32
2021-06-23T05:11:32
50,700,556
1
3
null
null
null
null
UTF-8
Python
false
false
454
py
# Find median of an Array of numbers def Median(ary): ary.sort() print(ary) if len(ary)%2!=0: mid=int(len(ary)/2) return ary[mid] else: idx1=int(len(ary)/2)-1 idx2=int(len(ary)/2) return (ary[idx1]+ary[idx2])/2 def main(): ary=[5, 89, 20, 64, 20, 45] print(Median(ary)) ary = [5, 89, 20, 64, 20, 45, 45, 23, 67, 32, 30] print(Median(ary)) if __name__=='__main__': main()
[ "kartiksayee@gmail.com" ]
kartiksayee@gmail.com
66598ec0257be4760b1dc38c8228fa5050235c13
6929a33a7259dad9b45192ca088a492085ed2953
/solutions/0062-unique-paths/unique-paths.py
1ac2c0fb3f2cb34e18806ccd977c68cdb4bb37c1
[]
no_license
moqi112358/leetcode
70366d29c474d19c43180fd4c282cc02c890af03
fab9433ff7f66d00023e3af271cf309b2d481722
refs/heads/master
2022-12-10T01:46:14.799231
2021-01-14T05:00:09
2021-01-14T05:00:09
218,163,960
3
0
null
2022-07-06T20:26:38
2019-10-28T23:26:47
Python
UTF-8
Python
false
false
1,747
py
# A robot is located at the top-left corner of a m x n grid (marked 'Start' in the diagram below). # # The robot can only move either down or right at any point in time. The robot is trying to reach the bottom-right corner of the grid (marked 'Finish' in the diagram below). # # How many possible unique paths are there? # #   # Example 1: # # # Input: m = 3, n = 7 # Output: 28 # # # Example 2: # # # Input: m = 3, n = 2 # Output: 3 # Explanation: # From the top-left corner, there are a total of 3 ways to reach the bottom-right corner: # 1. Right -> Down -> Down # 2. Down -> Down -> Right # 3. Down -> Right -> Down # # # Example 3: # # # Input: m = 7, n = 3 # Output: 28 # # # Example 4: # # # Input: m = 3, n = 3 # Output: 6 # # #   # Constraints: # # # 1 <= m, n <= 100 # It's guaranteed that the answer will be less than or equal to 2 * 109. # # class Solution: # def uniquePaths(self, m: int, n: int) -> int: # self.count = 0 # self.searchPath(m, n, 0, 0) # return self.count # def searchPath(self, m, n, r, c): # if r == m - 1 and c == n - 1: # self.count += 1 # dx = [0, 1] # dy = [1, 0] # for i in range(2): # x = r + dx[i] # y = c + dy[i] # if 0 <= x <= m - 1 and 0 <= y <= n - 1: # self.searchPath(m, n, x, y) # return def uniquePaths(self, m: int, n: int) -> int: res = [[0] * m for i in range(n)] for i in range(m): res[0][i] = 1 for i in range(n): res[i][0] = 1 for i in range(1, n): for j in range(1, m): res[i][j] = res[i-1][j] + res[i][j-1] return res[n-1][m-1]
[ "983028670@qq.com" ]
983028670@qq.com
8c987f3e7fcca91265fa510ed17ea01cb6f62bab
b2487a96bb865cfa0d1906c4e66a4aea9b613ce0
/pynfb/setup.py
fa813e333cb6ac1dadf94028ece66e8e86523568
[]
no_license
gurasog/nfb
1ccbbc7d507525cff65f5d5c756afd98ad86a7c6
9ff2c736c5d764f48d921bad3942c4db93390a5d
refs/heads/master
2022-12-14T23:06:56.523586
2020-09-07T07:36:01
2020-09-07T07:36:01
273,970,746
0
0
null
2020-06-21T19:30:38
2020-06-21T19:30:37
null
UTF-8
Python
false
false
164
py
from cx_Freeze import setup, Executable setup( name = "pynfb", version = "0.1", description = "Python NFB", executables = [Executable("main.py")] )
[ "smtnm@ya.ru" ]
smtnm@ya.ru
65ba12265ec9faffe3d18ec718e81ba5893e0a1e
d659810b24ebc6ae29a4d7fbb3b82294c860633a
/aliyun-python-sdk-unimkt/aliyunsdkunimkt/request/v20181212/ScanCodeNotificationRequest.py
5835e1a20aefefe670ace4df8bd671a5268d659a
[ "Apache-2.0" ]
permissive
leafcoder/aliyun-openapi-python-sdk
3dd874e620715173b6ccf7c34646d5cb8268da45
26b441ab37a5cda804de475fd5284bab699443f1
refs/heads/master
2023-07-31T23:22:35.642837
2021-09-17T07:49:51
2021-09-17T07:49:51
407,727,896
0
0
NOASSERTION
2021-09-18T01:56:10
2021-09-18T01:56:09
null
UTF-8
Python
false
false
5,343
py
# Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # # http://www.apache.org/licenses/LICENSE-2.0 # # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. from aliyunsdkcore.request import RpcRequest from aliyunsdkunimkt.endpoint import endpoint_data class ScanCodeNotificationRequest(RpcRequest): def __init__(self): RpcRequest.__init__(self, 'UniMkt', '2018-12-12', 'ScanCodeNotification') self.set_protocol_type('https') self.set_method('POST') if hasattr(self, "endpoint_map"): setattr(self, "endpoint_map", endpoint_data.getEndpointMap()) if hasattr(self, "endpoint_regional"): setattr(self, "endpoint_regional", endpoint_data.getEndpointRegional()) def get_RealCostAmount(self): return self.get_query_params().get('RealCostAmount') def set_RealCostAmount(self,RealCostAmount): self.add_query_param('RealCostAmount',RealCostAmount) def get_SalePrice(self): return self.get_query_params().get('SalePrice') def set_SalePrice(self,SalePrice): self.add_query_param('SalePrice',SalePrice) def get_CommodityId(self): return self.get_query_params().get('CommodityId') def set_CommodityId(self,CommodityId): self.add_query_param('CommodityId',CommodityId) def get_HolderId(self): return self.get_query_params().get('HolderId') def set_HolderId(self,HolderId): self.add_query_param('HolderId',HolderId) def get_DeviceType(self): return self.get_query_params().get('DeviceType') def set_DeviceType(self,DeviceType): self.add_query_param('DeviceType',DeviceType) def get_DeviceCode(self): return self.get_query_params().get('DeviceCode') def set_DeviceCode(self,DeviceCode): self.add_query_param('DeviceCode',DeviceCode) def get_ApplyPrice(self): return self.get_query_params().get('ApplyPrice') def set_ApplyPrice(self,ApplyPrice): self.add_query_param('ApplyPrice',ApplyPrice) def get_TaskId(self): return self.get_query_params().get('TaskId') def set_TaskId(self,TaskId): self.add_query_param('TaskId',TaskId) def get_OuterCode(self): return self.get_query_params().get('OuterCode') def set_OuterCode(self,OuterCode): self.add_query_param('OuterCode',OuterCode) def get_QueryStr(self): return self.get_query_params().get('QueryStr') def set_QueryStr(self,QueryStr): self.add_query_param('QueryStr',QueryStr) def get_Phase(self): return self.get_query_params().get('Phase') def set_Phase(self,Phase): self.add_query_param('Phase',Phase) def get_BizResult(self): return self.get_query_params().get('BizResult') def set_BizResult(self,BizResult): self.add_query_param('BizResult',BizResult) def get_TaskType(self): return self.get_query_params().get('TaskType') def set_TaskType(self,TaskType): self.add_query_param('TaskType',TaskType) def get_BrandUserId(self): return self.get_query_params().get('BrandUserId') def set_BrandUserId(self,BrandUserId): self.add_query_param('BrandUserId',BrandUserId) def get_Sex(self): return self.get_query_params().get('Sex') def set_Sex(self,Sex): self.add_query_param('Sex',Sex) def get_CostDetail(self): return self.get_query_params().get('CostDetail') def set_CostDetail(self,CostDetail): self.add_query_param('CostDetail',CostDetail) def get_ProxyUserId(self): return self.get_query_params().get('ProxyUserId') def set_ProxyUserId(self,ProxyUserId): self.add_query_param('ProxyUserId',ProxyUserId) def get_AlipayOpenId(self): return self.get_query_params().get('AlipayOpenId') def set_AlipayOpenId(self,AlipayOpenId): self.add_query_param('AlipayOpenId',AlipayOpenId) def get_BizType(self): return self.get_query_params().get('BizType') def set_BizType(self,BizType): self.add_query_param('BizType',BizType) def get_BrandNick(self): return self.get_query_params().get('BrandNick') def set_BrandNick(self,BrandNick): self.add_query_param('BrandNick',BrandNick) def get_V(self): return self.get_query_params().get('V') def set_V(self,V): self.add_query_param('V',V) def get_ChargeTag(self): return self.get_query_params().get('ChargeTag') def set_ChargeTag(self,ChargeTag): self.add_query_param('ChargeTag',ChargeTag) def get_Age(self): return self.get_query_params().get('Age') def set_Age(self,Age): self.add_query_param('Age',Age) def get_ChannelId(self): return self.get_query_params().get('ChannelId') def set_ChannelId(self,ChannelId): self.add_query_param('ChannelId',ChannelId) def get_Cid(self): return self.get_query_params().get('Cid') def set_Cid(self,Cid): self.add_query_param('Cid',Cid)
[ "sdk-team@alibabacloud.com" ]
sdk-team@alibabacloud.com
2bebb31050c1eea8eec0d868ef0ad1ce49749f7d
38bee274b237e508175be4c37bf357621ed50195
/pde/pdes/allen_cahn.py
f1a80a46cb539fbec959f00de2c5a1ab654a3641
[ "MIT" ]
permissive
anna-11/py-pde
26f0110266fdb21803e665447b1204bedb401d78
5b596af5f224e3ec2a7fbea8f87fab3896c19642
refs/heads/master
2023-06-17T09:13:34.026794
2021-07-09T12:10:36
2021-07-09T12:10:36
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,296
py
""" A Allen-Cahn equation .. codeauthor:: David Zwicker <david.zwicker@ds.mpg.de> """ from typing import Callable # @UnusedImport import numpy as np from ..fields import ScalarField from ..grids.boundaries.axes import BoundariesData from ..tools.docstrings import fill_in_docstring from ..tools.numba import jit, nb from .base import PDEBase, expr_prod class AllenCahnPDE(PDEBase): r"""A simple Allen-Cahn equation The mathematical definition is .. math:: \partial_t c = \gamma \nabla^2 c - c^3 + c where :math:`c` is a scalar field and :math:`\gamma` sets the interfacial width. """ explicit_time_dependence = False interface_width: float @fill_in_docstring def __init__(self, interface_width: float = 1, bc: BoundariesData = "natural"): """ Args: interface_width (float): The diffusivity of the described species bc: The boundary conditions applied to the field. {ARG_BOUNDARIES} """ super().__init__() self.interface_width = interface_width self.bc = bc @property def expression(self) -> str: """str: the expression of the right hand side of this PDE""" return f"{expr_prod(self.interface_width, 'laplace(c)')} - c**3 + c" def evolution_rate( # type: ignore self, state: ScalarField, t: float = 0, ) -> ScalarField: """evaluate the right hand side of the PDE Args: state (:class:`~pde.fields.ScalarField`): The scalar field describing the concentration distribution t (float): The current time point Returns: :class:`~pde.fields.ScalarField`: Scalar field describing the evolution rate of the PDE """ assert isinstance(state, ScalarField), "`state` must be ScalarField" laplace = state.laplace(bc=self.bc, label="evolution rate") return self.interface_width * laplace - state ** 3 + state # type: ignore def _make_pde_rhs_numba( # type: ignore self, state: ScalarField ) -> Callable[[np.ndarray, float], np.ndarray]: """create a compiled function evaluating the right hand side of the PDE Args: state (:class:`~pde.fields.ScalarField`): An example for the state defining the grid and data types Returns: A function with signature `(state_data, t)`, which can be called with an instance of :class:`~numpy.ndarray` of the state data and the time to obtained an instance of :class:`~numpy.ndarray` giving the evolution rate. """ shape = state.grid.shape arr_type = nb.typeof(np.empty(shape, dtype=state.data.dtype)) signature = arr_type(arr_type, nb.double) interface_width = self.interface_width laplace = state.grid.get_operator("laplace", bc=self.bc) @jit(signature) def pde_rhs(state_data: np.ndarray, t: float) -> np.ndarray: """compiled helper function evaluating right hand side""" return interface_width * laplace(state_data) - state_data ** 3 + state_data # type: ignore return pde_rhs # type: ignore
[ "david.zwicker@ds.mpg.de" ]
david.zwicker@ds.mpg.de
c2568db3ed9af7c59ba95dc30e283b4abd83d2a1
e9adf4bb294b22add02c997f750e36b6fea23cdc
/nw/nw_logic/nw_rules_bank.py
432c3a0479a0241af1dc0f9ec9c891ae0851037a
[ "BSD-3-Clause" ]
permissive
bairoliyaprem/python-rules
5aa70dbf6d9efde0d95dda4b6397e36788c22cfa
ca7ab0addf60179ea61fddad6cd65c77a6792de1
refs/heads/master
2022-12-17T22:15:47.745660
2020-09-29T00:39:03
2020-09-29T00:39:03
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,863
py
import sqlalchemy from sqlalchemy_utils import get_mapper from python_rules.exec_row_logic.logic_row import LogicRow from python_rules.rule import Rule from python_rules.rule_bank.rule_bank import RuleBank from nw.nw_logic import models from nw.nw_logic.models import Customer, OrderDetail, Product, Order, Employee def activate_basic_check_credit_rules(): """ Issues function calls to activate check credit rules, below. These rules are executed not now, but on commits Order is irrelevant - determined by system based on dependency analysis Their inclusion in classes is for doc / convenience, no semantics These rules apply to all transactions (automatic re-use), eg. * place order * change Order Detail product, quantity * add/delete Order Detail * ship / unship order * delete order * move order to new customer, etc """ def units_in_stock(row: Product, old_row: Product, logic_row: LogicRow): result = row.UnitsInStock - (row.UnitsShipped - old_row.UnitsShipped) return result def congratulate_sales_rep(row: Order, old_row: Order, logic_row: LogicRow): if logic_row.ins_upd_dlt == "ins" or True: # logic engine fills parents for insert sales_rep = row.SalesRep # type : Employee if sales_rep is None: logic_row.log("no salesrep for this order") else: logic_row.log(f'Hi, {sales_rep.Manager.FirstName}, congratulate {sales_rep.FirstName} on their new order') Rule.constraint(validate=Customer, as_condition=lambda row: row.Balance <= row.CreditLimit, error_msg="balance ({row.Balance}) exceeds credit ({row.CreditLimit})") Rule.sum(derive=Customer.Balance, as_sum_of=Order.AmountTotal, where=lambda row: row.ShippedDate is None) # *not* a sql select sum... Rule.sum(derive=Order.AmountTotal, as_sum_of=OrderDetail.Amount) Rule.formula(derive=OrderDetail.Amount, as_expression=lambda row: row.UnitPrice * row.Quantity) Rule.copy(derive=OrderDetail.UnitPrice, from_parent=Product.UnitPrice) Rule.formula(derive=OrderDetail.ShippedDate, as_exp="row.OrderHeader.ShippedDate") Rule.sum(derive=Product.UnitsShipped, as_sum_of=OrderDetail.Quantity, where="row.ShippedDate is not None") Rule.formula(derive=Product.UnitsInStock, calling=units_in_stock) Rule.commit_row_event(on_class=Order, calling=congratulate_sales_rep) Rule.count(derive=Customer.UnpaidOrderCount, as_count_of=Order, where=lambda row: row.ShippedDate is None) # *not* a sql select sum... Rule.count(derive=Customer.OrderCount, as_count_of=Order) class InvokePythonFunctions: # use functions for more complex rules, type checking, etc (not used) @staticmethod def load_rules(self): def my_early_event(row, old_row, logic_row): logic_row.log("early event for *all* tables - good breakpoint, time/date stamping, etc") def check_balance(row: Customer, old_row, logic_row) -> bool: """ Not used... illustrate function alternative (e.g., more complex if/else logic) specify rule with `calling=check_balance` (instead of as_condition) """ return row.Balance <= row.CreditLimit def compute_amount(row: OrderDetail, old_row, logic_row): return row.UnitPrice * row.Quantity Rule.formula(derive="OrderDetail.Amount", calling=compute_amount) Rule.formula(derive="OrderDetail.Amount", calling=lambda Customer: Customer.Quantity * Customer.UnitPrice) Rule.early_row_event(on_class="*", calling=my_early_event) # just for debug Rule.constraint(validate="Customer", calling=check_balance, error_msg="balance ({row.Balance}) exceeds credit ({row.CreditLimit})") class DependencyGraphTests: """Not loaded""" def not_loaded(self): Rule.formula(derive="Tbl.ColA", # or, calling=compute_amount) as_exp="row.ColB + row.ColC") Rule.formula(derive="Tbl.ColB", # or, calling=compute_amount) as_exp="row.ColC") Rule.formula(derive="Tbl.ColC", # or, calling=compute_amount) as_exp="row.ColD") Rule.formula(derive="Tbl.ColD", # or, calling=compute_amount) as_exp="row.ColE") Rule.formula(derive="Tbl.ColE", # or, calling=compute_amount) as_exp="xxx") class UnusedTests: """Not loaded""" def not_loaded(self): Rule.constraint(validate="AbUser", # table is ab_user calling=lambda row: row.username != "no_name") Rule.count(derive=Customer.OrderCount, as_count_of=Order, where="ShippedDate not None")
[ "valjhuber@gmail.com" ]
valjhuber@gmail.com
7ca0ee05969af630f9b2b5a8871790f4991b5c08
d55f3f715c00bcbd60badb3a31696a1a629600e2
/students/maks/9/site2/page/management/commands/hello.py
014dfc85a65d4e56ef7b7b974c84bd8f9de0f150
[]
no_license
zdimon/wezom-python-course
ea0adaa54444f6deaca81ce54ee8334297f2cd1a
5b87892102e4eb77a4c12924d2d71716b9cce721
refs/heads/master
2023-01-29T02:22:54.220880
2020-12-05T11:27:48
2020-12-05T11:27:48
302,864,112
0
0
null
null
null
null
UTF-8
Python
false
false
523
py
from django.core.management.base import BaseCommand, CommandError from page.models import Page class Command(BaseCommand): def handle(self, *args, **options): print('Hello command!!!') Page.objects.all().delete() page1 = Page() page1.title = 'Index page' page1.content = 'content content' page1.save() page2 = Page() page2.title = 'Index page 2' page2.content = 'content content 2' page2.save()
[ "you@example.com" ]
you@example.com
4df98b423cb6c082c08b54fdbf134f6143af4b37
cb6f314d0c8f1a943718caa46b47bbe5ef9908f5
/test/drawings/draw_hmab.py
21c985c2897367265b4cd012a1bfb34b55a9bbce
[ "MIT" ]
permissive
yunx-z/soln-ml
d702d5afef7d9204c89bdcba44e26f89a68b8923
f97c991c2ad287e8d295d3058b4a9b1fd50d847b
refs/heads/master
2022-12-26T18:00:51.457311
2020-09-13T08:31:58
2020-09-13T08:31:58
296,199,472
0
0
MIT
2020-09-17T02:37:20
2020-09-17T02:37:20
null
UTF-8
Python
false
false
1,574
py
import os import sys import time import pickle import argparse import numpy as np import autosklearn.classification sys.path.append(os.getcwd()) parser = argparse.ArgumentParser() dataset_set = 'diabetes,spectf,credit,ionosphere,lymphography,pc4,' \ 'messidor_features,winequality_red,winequality_white,splice,spambase,amazon_employee' parser.add_argument('--datasets', type=str, default=dataset_set) parser.add_argument('--mth', choices=['ours', 'ausk'], default='ours') parser.add_argument('--seed', type=int, default=1) parser.add_argument('--algo_num', type=int, default=8) parser.add_argument('--trial_num', type=int, default=100) project_dir = './' def plot(mth, dataset, algo_num, trial_num, seed): if mth == 'ours': save_path = project_dir + 'data/hmab_%s_%d_%d_%d.pkl' % \ (dataset, trial_num, algo_num, seed) else: save_path = project_dir + 'data/ausk_%s_%d.pkl' % (dataset, algo_num) with open(save_path, 'rb') as f: result = pickle.load(f) print('Best validation accuracy: %.4f' % np.max(result[0])) print('Final Rewards', result[0]) print('Time records', result[1]) print('Action Sequence', result[2]) print('-' * 30) if __name__ == "__main__": args = parser.parse_args() dataset_str = args.datasets dataset_list = list() if dataset_str == 'all': dataset_list = dataset_set else: dataset_list = dataset_str.split(',') for dataset in dataset_list: plot(args.mth, dataset, args.algo_num, args.trial_num, args.seed)
[ "1225646303@qq.com" ]
1225646303@qq.com
dce5217f537b1368245960c5d19597735d145b4a
b6fc54cff7037f5e4ef26cb4a645d5ea5a6fecdf
/001146StepikPyBegin/Stepik001146PyBeginсh07p01st01T01_for_20200420.py
e9b62e163ccebe48ce59797d086fef7a96ccb3fa
[ "Apache-2.0" ]
permissive
SafonovMikhail/python_000577
5483eaf2f7c73bc619ce1f5de67d8d689d2e7dd4
f2dccac82a37df430c4eb7425b5d084d83520409
refs/heads/master
2022-12-08T10:53:57.202746
2022-12-07T09:09:51
2022-12-07T09:09:51
204,713,341
0
0
null
null
null
null
UTF-8
Python
false
false
203
py
# реализация множественного ввода for i in range(5): num1 = int(input()) print("Квадрат числа равен:", num1 ** 2) print("Цикл завершен")
[ "ms33@inbox.ru" ]
ms33@inbox.ru
884a0b600ab0d1c224f04dbc9a3f59c3b47ca4f1
fcd965c9333ee328ec51bc41f5bc0300cc06dc33
/Coding Patterns/Fast & Slow Pointers/Palindrome LinkedList.py
8e1dc912eb07d240edfe2d100acd686917699403
[]
no_license
henrylin2008/Coding_Problems
699bb345481c14dc3faa8bab439776c7070a1cb0
281067e872f73a27f76ae10ab0f1564916bddd28
refs/heads/master
2023-01-11T11:55:47.936163
2022-12-24T07:50:17
2022-12-24T07:50:17
170,151,972
2
0
null
null
null
null
UTF-8
Python
false
false
3,389
py
# Problem Challenge 1: Palindrome LinkedList (medium) # https://designgurus.org/path-player?courseid=grokking-the-coding-interview&unit=grokking-the-coding-interview_1628743582805_17Unit # Problem Statement # # Given the head of a Singly LinkedList, write a method to check if the LinkedList is a palindrome or not. # # Your algorithm should use constant space and the input LinkedList should be in the original form once the algorithm # is finished. The algorithm should have O(N)O(N) time complexity where ‘N’ is the number of nodes in the LinkedList. # # Example 1: # Input: 2 -> 4 -> 6 -> 4 -> 2 -> null # Output: true # # Example 2: # Input: 2 -> 4 -> 6 -> 4 -> 2 -> 2 -> null # Output: false # Solution # # As we know, a palindrome LinkedList will have nodes values that read the same backward or forward. This means that # if we divide the LinkedList into two halves, the node values of the first half in the forward direction should be # similar to the node values of the second half in the backward direction. As we have been given a Singly LinkedList, # we can’t move in the backward direction. To handle this, we will perform the following steps: # 1. We can use the Fast & Slow pointers method similar to Middle of the LinkedList to find the middle node of the # LinkedList. # 2. Once we have the middle of the LinkedList, we will reverse the second half. # 3. Then, we will compare the first half with the reversed second half to see if the LinkedList represents a # palindrome. # 4. Finally, we will reverse the second half of the LinkedList again to revert and bring the LinkedList back to its # original form. class Node: def __init__(self, value, next=None): self.value = value self.next = next def is_palindromic_linked_list(head): if head is None or head.next is None: return True # find middle of the LinkedList slow, fast = head, head while fast is not None and fast.next is not None: slow = slow.next fast = fast.next.next head_second_half = reverse(slow) # reverse the second half # store the head of reversed part to revert back later copy_head_second_half = head_second_half # compare the first and the second half while head is not None and head_second_half is not None: if head.value != head_second_half.value: break # not a palindrome head = head.next head_second_half = head_second_half.next reverse(copy_head_second_half) # revert the reverse of the second half if head is None or head_second_half is None: # if both halves match return True return False def reverse(head): prev = None while head is not None: next = head.next head.next = prev prev = head head = next return prev def main(): head = Node(2) head.next = Node(4) head.next.next = Node(6) head.next.next.next = Node(4) head.next.next.next.next = Node(2) print("Is palindrome: " + str(is_palindromic_linked_list(head))) head.next.next.next.next.next = Node(2) print("Is palindrome: " + str(is_palindromic_linked_list(head))) main() # Time Complexity # The above algorithm will have a time complexity of O(N) where ‘N’ is the number of nodes in the LinkedList. # # Space Complexity # The algorithm runs in constant space O(1).
[ "henrylin2008@yahoo.com" ]
henrylin2008@yahoo.com
dcb3980fb3389f1967bc679b30b0ca7aa6b476c9
08d316151302f7ba4ae841c15b7adfe4e348ddf1
/reviewboard/hostingsvcs/tests/test_fogbugz.py
97cccab17520fd5b84e7a49e97c38afd16cd410c
[ "MIT" ]
permissive
LloydFinch/reviewboard
aa8cd21fac359d49b3dfc5a68c42b857c0c04bd8
563c1e8d4dfd860f372281dc0f380a0809f6ae15
refs/heads/master
2020-08-10T20:02:32.204351
2019-10-02T20:46:08
2019-10-02T20:46:08
214,411,166
2
0
MIT
2019-10-11T10:44:55
2019-10-11T10:44:54
null
UTF-8
Python
false
false
952
py
"""Unit tests for the FogBugz hosting service.""" from __future__ import unicode_literals from reviewboard.hostingsvcs.testing import HostingServiceTestCase class FogBugzTests(HostingServiceTestCase): """Unit tests for the FogBugz hosting service.""" service_name = 'fogbugz' fixtures = ['test_scmtools'] def test_service_support(self): """Testing FogBugz service support capabilities""" self.assertTrue(self.service_class.supports_bug_trackers) self.assertFalse(self.service_class.supports_repositories) def test_get_bug_tracker_field(self): """Testing FogBugz.get_bug_tracker_field""" self.assertFalse( self.service_class.get_bug_tracker_requires_username()) self.assertEqual( self.service_class.get_bug_tracker_field(None, { 'fogbugz_account_domain': 'mydomain', }), 'https://mydomain.fogbugz.com/f/cases/%s')
[ "christian@beanbaginc.com" ]
christian@beanbaginc.com
a3ed4c22c27ef24b07b4ce76b60d632ea251f1f0
de24f83a5e3768a2638ebcf13cbe717e75740168
/moodledata/vpl_data/212/usersdata/265/87081/submittedfiles/av1_3.py
1d4a47463f8274d69023d764799f71ec143c55a0
[]
no_license
rafaelperazzo/programacao-web
95643423a35c44613b0f64bed05bd34780fe2436
170dd5440afb9ee68a973f3de13a99aa4c735d79
refs/heads/master
2021-01-12T14:06:25.773146
2017-12-22T16:05:45
2017-12-22T16:05:45
69,566,344
0
0
null
null
null
null
UTF-8
Python
false
false
365
py
# -*- coding: utf-8 -*- import math a = int(input('digite o valor de a: ')) b = int(input('digite o valor de b: ')) c = int(input('digite o valor de c: ')) i=2 j=2 k=2 while i<=a: if a%i==0: divisor_a=i i=1+i print(i) while j<=b: if b%j==0: divisor_b=j j=1+j print(j) while k<=c: if c%k==0: divisor_c=k k=1+k print(k)
[ "rafael.mota@ufca.edu.br" ]
rafael.mota@ufca.edu.br
1849bac316628694ea87ff00774d2816a84a04ce
6e4e395988c641856aa13aa3b68db838f0d47cc0
/Trying out DB-API.py
7a7c01c86e3f781089783e9aaf81459aec431027
[ "MIT" ]
permissive
fatih-iver/Intro-to-Relational-Databases
2669d060fd2f9c8884e936e541373eecfbe8634e
28528132378436d6dd1f1bdec96d1e7e285b4e4d
refs/heads/master
2020-03-16T11:26:54.090473
2018-05-09T21:32:38
2018-05-09T21:32:38
132,648,541
0
0
null
null
null
null
UTF-8
Python
false
false
667
py
# To see how the various functions in the DB-API work, take a look at this code, # then the results that it prints when you press "Test Run". # # Then modify this code so that the student records are fetched in sorted order # by student's name. # import sqlite3 # Fetch some student records from the database. db = sqlite3.connect("students") c = db.cursor() query = "select name, id from students order by name;" c.execute(query) rows = c.fetchall() # First, what data structure did we get? print("Row data:") print(rows) # And let's loop over it too: print() print("Student names:") for row in rows: print(" ", row[0]) db.close()
[ "noreply@github.com" ]
fatih-iver.noreply@github.com
fa1d83149490bbaaf6e2373771de847c30a4191b
10326daa458342fd992f4bd2f9c63c9469ae5a11
/.graal-git-repo
419b5f2ded2ecc3adce67832c5ddb0e5577ca1db
[ "MIT", "CC-BY-SA-4.0" ]
permissive
charig/black-diamonds
b08c1d9dab528a9c9c935414e6af4faf2d5773fb
0fd46caeec41f57b621143be1e4bd22d5a9a7348
refs/heads/master
2020-04-12T20:04:36.490970
2018-09-22T13:39:03
2018-09-22T13:39:03
162,725,323
0
0
MIT
2018-12-21T14:47:34
2018-12-21T14:47:34
null
UTF-8
Python
false
false
2,522
#!/usr/bin/python # This is a self-updating script encoding the Graal+Truffle and mx repos and # revision used for testing on CI. We don't use submodules to avoid a hard # dependency that might bloat the repository of users of BlackDiamonds # # To checkout the repos, at the specified version for this version of the code, # run `./graal-git-repo checkout` # To update this script, so its revisions point to the latest versions of the # configured repos, run `./graal-git-repo update-script-revs` import sys import os # We use the following repositories GRAAL_REPO_URL = "https://github.com/smarr/truffle.git" MX_REPO_URL = "https://github.com/graalvm/mx.git" # And these are the repo revisions we test against GRAAL_REPO_REV = "a9fba6a775ffc60a90959d2eff4e66d15e9867a9" MX_REPO_REV = "5bc7f83b9d66a31259b90933fcd0aa64d38b8d1e" def update(lines, var, val): for idx, line in enumerate(lines): if line.startswith(var): print("Updating " + var + " to " + val) lines[idx] = var.ljust(15) + '= "' + val + '"\n' break def run(cmd): print(cmd) return os.popen(cmd).read() if len(sys.argv) == 1: print("To checkout the Graal+Truffle and MX dependencies use:") print(" " + __file__ + " checkout") print("To update the dependencies in this script use:") print(" " + __file__ + " update-script-revs") quit() if sys.argv[1] == "update-script-revs": graal_head_data = run("git ls-remote " + GRAAL_REPO_URL + " HEAD") graal_head_rev = graal_head_data.split("\t")[0] mx_head_data = run("git ls-remote " + MX_REPO_URL + " HEAD") mx_head_rev = mx_head_data.split("\t")[0] with open(__file__, 'r') as script_file: content = script_file.readlines() update(content, 'GRAAL_REPO_REV', graal_head_rev) update(content, 'MX_REPO_REV', mx_head_rev) with open(__file__, 'w') as script_file: script_file.writelines(content) def update_repo(folder, repo, rev): folder = os.path.realpath(folder) if not os.path.isdir(folder): print("cloning " + repo) print(run("git clone --depth 5000 " + repo + " " + folder)) run("git --git-dir=" + folder + "/.git --work-tree=" + folder + " fetch --depth 5000") print(run("git --git-dir=" + folder + "/.git --work-tree=" + folder + " reset --hard " + rev)) if sys.argv[1] == "checkout": update_repo("graal", GRAAL_REPO_URL, GRAAL_REPO_REV) update_repo("mx", MX_REPO_URL, MX_REPO_REV)
[ "git@stefan-marr.de" ]
git@stefan-marr.de
967c4163dd3b2b4b3fccef187338d4d020e0e693
7c63a96fad4257f4959ffeba0868059fc96566fb
/py/m_lutz-programming_python-4_ed/code/ch_01/step_06/01-cgi_basics/cgi-bin/cgi101.py
73cd3dab161b31965720666610c168715c174345
[ "MIT" ]
permissive
ordinary-developer/education
b426148f5690f48e0ed4853adfc3740bd038b72c
526e5cf86f90eab68063bb7c75744226f2c54b8d
refs/heads/master
2023-08-31T14:42:37.237690
2023-08-30T18:15:18
2023-08-30T18:15:18
91,232,306
8
0
null
null
null
null
UTF-8
Python
false
false
270
py
#!/usr/local/bin/python3 import cgi form = cgi.FieldStorage() print('Content-type: text/html\n') print('<title>Reply Page</title>') if not 'user' in form: print('<h1>Who are you?</h1>') else: print('<h1>Hello <i>%s</i>!</h1>' % cgi.escape(form['user'].value))
[ "merely.ordinary.developer@gmail.com" ]
merely.ordinary.developer@gmail.com