blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
eb02dc1e6b7e567cad140a6e64b4ffcbd14c3557
44f3a3b7541c99554cbd94b015a8c25593f3e95e
/email/pyse_mail/public/login.py
d317bbd7ab8d43b189f8f4b3b87307630db4bbd1
[]
no_license
reblues/study
adc2d255c4c5990fcfcd4d8691625057c95357d8
7ea2854c778279d40d1dd3df362f463774f5971b
refs/heads/master
2020-12-31T05:24:13.440027
2016-06-01T06:28:33
2016-06-01T06:28:33
58,917,230
0
0
null
null
null
null
UTF-8
Python
false
false
907
py
#!/user/bin/env python # -*- coding:utf-8 -*- from selenium import webdriver from selenium.webdriver.common.action_chains import ActionChains from selenium.webdriver.common.keys import Keys from selenium.webdriver.support.ui import WebDriverWait from selenium.common.exceptions import NoSuchElementException import time,unittest def Login(self): driver = self.driver driver.maximize_window() target = driver.find_element_by_id("lbNormal") ActionChains(driver).move_to_element(target).perform() time.sleep(2) driver.find_element_by_id("idInput").send_keys("18710924623") driver.find_element_by_id("pwdInput").send_keys("yang9351yu") driver.find_element_by_id("loginBtn").click() time.sleep(2) print "登录成功" def Logout(self): driver = self.driver time.sleep(2) aim = driver.find_element_by_link_text(u"退出") print aim.text aim.click()
[ "534948651@qq.com" ]
534948651@qq.com
9514286077c40b1598552cdc24d2d2d31844d5fe
34ed92a9593746ccbcb1a02630be1370e8524f98
/lib/pints/pints/tests/test_mcmc_relativistic.py
1fb0e2abb531defd9c0d3b86dccf543b66d3e108
[ "LicenseRef-scancode-unknown-license-reference", "BSD-3-Clause" ]
permissive
HOLL95/Cytochrome_SV
87b7a680ed59681230f79e1de617621680ea0fa0
d02b3469f3ee5a4c85d756053bc87651093abea1
refs/heads/master
2022-08-01T05:58:16.161510
2021-02-01T16:09:31
2021-02-01T16:09:31
249,424,867
0
0
null
2022-06-22T04:09:11
2020-03-23T12:29:29
Jupyter Notebook
UTF-8
Python
false
false
6,142
py
#!/usr/bin/env python3 # # Tests the basic methods of the Relativistic MCMC routine. # # This file is part of PINTS. # Copyright (c) 2017-2019, University of Oxford. # For licensing information, see the LICENSE file distributed with the PINTS # software package. # import unittest import numpy as np import pints import pints.toy from shared import StreamCapture class TestRelativisticMCMC(unittest.TestCase): """ Tests the basic methods of the Relativistic MCMC routine. """ def test_method(self): # Create log pdf log_pdf = pints.toy.GaussianLogPDF([5, 5], [[4, 1], [1, 3]]) # Create mcmc x0 = np.array([2, 2]) sigma = [[3, 0], [0, 3]] mcmc = pints.RelativisticMCMC(x0, sigma) # This method needs sensitivities self.assertTrue(mcmc.needs_sensitivities()) # Set number of leapfrog steps ifrog = 10 mcmc.set_leapfrog_steps(ifrog) # Perform short run chain = [] for i in range(100 * ifrog): x = mcmc.ask() fx, gr = log_pdf.evaluateS1(x) sample = mcmc.tell((fx, gr)) if i >= 50 * ifrog and sample is not None: chain.append(sample) if np.all(sample == x): self.assertEqual(mcmc.current_log_pdf(), fx) chain = np.array(chain) self.assertEqual(chain.shape[0], 50) self.assertEqual(chain.shape[1], len(x0)) def test_logging(self): """ Test logging includes name and custom fields. """ log_pdf = pints.toy.GaussianLogPDF([5, 5], [[4, 1], [1, 3]]) x0 = [np.array([2, 2]), np.array([8, 8])] mcmc = pints.MCMCController( log_pdf, 2, x0, method=pints.RelativisticMCMC) mcmc.set_max_iterations(5) with StreamCapture() as c: mcmc.run() text = c.text() self.assertIn('Relativistic MCMC', text) self.assertIn(' Accept.', text) def test_flow(self): log_pdf = pints.toy.GaussianLogPDF([5, 5], [[4, 1], [1, 3]]) x0 = np.array([2, 2]) # Test initial proposal is first point mcmc = pints.RelativisticMCMC(x0) self.assertTrue(np.all(mcmc.ask() == mcmc._x0)) # Repeated asks self.assertRaises(RuntimeError, mcmc.ask) # Tell without ask mcmc = pints.RelativisticMCMC(x0) self.assertRaises(RuntimeError, mcmc.tell, 0) # Repeated tells should fail x = mcmc.ask() mcmc.tell(log_pdf.evaluateS1(x)) self.assertRaises(RuntimeError, mcmc.tell, log_pdf.evaluateS1(x)) # Bad starting point mcmc = pints.RelativisticMCMC(x0) mcmc.ask() self.assertRaises( ValueError, mcmc.tell, (float('-inf'), np.array([1, 1]))) def test_kinetic_energy(self): """ Tests kinetic energy values and derivatives """ x0 = np.array([2, 2]) model = pints.RelativisticMCMC(x0) model.ask() # kinetic energy mc2 = 100.0 momentum = [1.0, 2.0] squared = np.sum(np.array(momentum)**2) ke1 = mc2 * (squared / mc2 + 1.0)**0.5 ke2 = model._kinetic_energy(momentum) self.assertEqual(ke1, ke2) c = 1.0 m = 1.0 mc2 = m * c**2 squared = np.sum(np.array(momentum)**2) ke1 = mc2 * (squared / mc2 + 1.0)**0.5 model = pints.RelativisticMCMC(x0) model.set_speed_of_light(c) model.ask() ke2 = model._kinetic_energy(momentum) self.assertEqual(ke1, ke2) def test_set_hyper_parameters(self): """ Tests the parameter interface for this sampler. """ x0 = np.array([2, 2]) mcmc = pints.RelativisticMCMC(x0) # Test leapfrog parameters n = mcmc.leapfrog_steps() d = mcmc.leapfrog_step_size() self.assertIsInstance(n, int) self.assertTrue(len(d) == mcmc._n_parameters) mcmc.set_leapfrog_steps(n + 1) self.assertEqual(mcmc.leapfrog_steps(), n + 1) self.assertRaises(ValueError, mcmc.set_leapfrog_steps, 0) mcmc.set_leapfrog_step_size(0.5) self.assertEqual(mcmc.leapfrog_step_size()[0], 0.5) self.assertRaises(ValueError, mcmc.set_leapfrog_step_size, -1) self.assertEqual(mcmc.n_hyper_parameters(), 4) mcmc.set_hyper_parameters([n + 2, 2, 0.4, 2.3]) self.assertEqual(mcmc.leapfrog_steps(), n + 2) self.assertEqual(mcmc.leapfrog_step_size()[0], 2) self.assertEqual(mcmc.mass(), 0.4) self.assertEqual(mcmc.speed_of_light(), 2.3) mcmc.set_epsilon(0.4) self.assertEqual(mcmc.epsilon(), 0.4) self.assertRaises(ValueError, mcmc.set_epsilon, -0.1) mcmc.set_leapfrog_step_size(1) self.assertEqual(len(mcmc.scaled_epsilon()), 2) self.assertEqual(mcmc.scaled_epsilon()[0], 0.4) self.assertEqual(len(mcmc.divergent_iterations()), 0) self.assertRaises(ValueError, mcmc.set_leapfrog_step_size, [1, 2, 3]) mcmc.set_leapfrog_step_size([1.5, 3]) self.assertEqual(mcmc.leapfrog_step_size()[0], 1.5) self.assertEqual(mcmc.leapfrog_step_size()[1], 3) c = 3.5 mcmc.set_speed_of_light(c) self.assertEqual(mcmc.speed_of_light(), c) self.assertRaises(ValueError, mcmc.set_speed_of_light, -0.1) m = 2.9 mcmc.set_mass(m) self.assertEqual(mcmc.mass(), m) self.assertRaises(ValueError, mcmc.set_mass, -1.8) self.assertRaises(ValueError, mcmc.set_mass, [1, 3]) def test_other_setters(self): # Tests other setters and getters. x0 = np.array([2, 2]) mcmc = pints.RelativisticMCMC(x0) self.assertRaises(ValueError, mcmc.set_hamiltonian_threshold, -0.3) threshold1 = mcmc.hamiltonian_threshold() self.assertEqual(threshold1, 10**3) threshold2 = 10 mcmc.set_hamiltonian_threshold(threshold2) self.assertEqual(mcmc.hamiltonian_threshold(), threshold2) if __name__ == '__main__': unittest.main()
[ "henney@localhost.localdomain" ]
henney@localhost.localdomain
030af696a1ebdd2d98a56cc9345bfe20f5099896
67ceb35320d3d02867350bc6d460ae391e0324e8
/practice/hard/0675-Cut_Trees_for_Golf_Event.py
e91dcd1441c759908435b4cb1b2766949823a97b
[]
no_license
mattjp/leetcode
fb11cf6016aef46843eaf0b55314e88ccd87c91a
88ccd910dfdb0e6ca6a70fa2d37906c31f4b3d70
refs/heads/master
2023-01-22T20:40:48.104388
2022-12-26T22:03:02
2022-12-26T22:03:02
184,347,356
0
0
null
null
null
null
UTF-8
Python
false
false
1,504
py
class Solution: def cutOffTree(self, forest: List[List[int]]) -> int: """ 0. while there are trees to cut down 1. walk to coordinates of next tree; cut down - do BFS dummy 2. if tree is unreachable - return """ from collections import deque from sortedcontainers import SortedDict def go_to_tree(grid, i, j, tree) -> int: queue = deque([(i, j, 0)]) # (i, j, steps) visited = set() while queue: row, col, steps = queue.popleft() if (row, col) == tree: return steps for r,c in [(1,0), (-1,0), (0,1), (0,-1)]: new_row, new_col = row+r, col+c if ( new_row < len(grid) and new_col < len(grid[0]) and new_row > -1 and new_col > -1 and (new_row, new_col) not in visited and grid[new_row][new_col] != 0 ): if (new_row, new_col) == tree: return steps+1 visited.add((new_row, new_col)) queue.append((new_row, new_col, steps+1)) return None trees = SortedDict() for i in range(len(forest)): for j in range(len(forest[i])): if forest[i][j] > 1: trees[forest[i][j]] = (i,j) total_steps = 0 i = j = 0 for h,tree in trees.items(): steps = go_to_tree(forest, i, j, tree) if steps == None: return -1 total_steps += steps i,j = tree return total_steps
[ "noreply@github.com" ]
noreply@github.com
9bff11e3a8633333af71b3cc5a2bc2241e5e3ec0
68c182cbb167ec6870ec1a301958e71ce8f9bcbb
/test/functional/p2p_permissions.py
d59b0acadab68cf792b145888a6743bf9ce0b48e
[ "MIT" ]
permissive
megamcloud/umkoin
de10e9bbe0afbdc7210db56e41f823a0805283be
3e0d7a48f459ff09f0b9e02c3ed30563670009c8
refs/heads/master
2022-05-30T00:18:10.962521
2020-04-26T08:21:01
2020-04-26T08:21:01
null
0
0
null
null
null
null
UTF-8
Python
false
false
6,694
py
#!/usr/bin/env python3 # Copyright (c) 2015-2020 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. """Test p2p permission message. Test that permissions are correctly calculated and applied """ from test_framework.address import ADDRESS_BCRT1_P2WSH_OP_TRUE from test_framework.messages import ( CTransaction, CTxInWitness, FromHex, ) from test_framework.mininode import P2PDataStore from test_framework.script import ( CScript, OP_TRUE, ) from test_framework.test_node import ErrorMatch from test_framework.test_framework import UmkoinTestFramework from test_framework.util import ( assert_equal, connect_nodes, p2p_port, wait_until, ) class P2PPermissionsTests(UmkoinTestFramework): def set_test_params(self): self.num_nodes = 2 self.setup_clean_chain = True def run_test(self): self.check_tx_relay() self.checkpermission( # default permissions (no specific permissions) ["-whitelist=127.0.0.1"], ["relay", "noban", "mempool"], True) self.checkpermission( # relay permission removed (no specific permissions) ["-whitelist=127.0.0.1", "-whitelistrelay=0"], ["noban", "mempool"], True) self.checkpermission( # forcerelay and relay permission added # Legacy parameter interaction which set whitelistrelay to true # if whitelistforcerelay is true ["-whitelist=127.0.0.1", "-whitelistforcerelay"], ["forcerelay", "relay", "noban", "mempool"], True) # Let's make sure permissions are merged correctly # For this, we need to use whitebind instead of bind # by modifying the configuration file. ip_port = "127.0.0.1:{}".format(p2p_port(1)) self.replaceinconfig(1, "bind=127.0.0.1", "whitebind=bloomfilter,forcerelay@" + ip_port) self.checkpermission( ["-whitelist=noban@127.0.0.1"], # Check parameter interaction forcerelay should activate relay ["noban", "bloomfilter", "forcerelay", "relay"], False) self.replaceinconfig(1, "whitebind=bloomfilter,forcerelay@" + ip_port, "bind=127.0.0.1") self.checkpermission( # legacy whitelistrelay should be ignored ["-whitelist=noban,mempool@127.0.0.1", "-whitelistrelay"], ["noban", "mempool"], False) self.checkpermission( # legacy whitelistforcerelay should be ignored ["-whitelist=noban,mempool@127.0.0.1", "-whitelistforcerelay"], ["noban", "mempool"], False) self.checkpermission( # missing mempool permission to be considered legacy whitelisted ["-whitelist=noban@127.0.0.1"], ["noban"], False) self.checkpermission( # all permission added ["-whitelist=all@127.0.0.1"], ["forcerelay", "noban", "mempool", "bloomfilter", "relay"], False) self.stop_node(1) self.nodes[1].assert_start_raises_init_error(["-whitelist=oopsie@127.0.0.1"], "Invalid P2P permission", match=ErrorMatch.PARTIAL_REGEX) self.nodes[1].assert_start_raises_init_error(["-whitelist=noban@127.0.0.1:230"], "Invalid netmask specified in", match=ErrorMatch.PARTIAL_REGEX) self.nodes[1].assert_start_raises_init_error(["-whitebind=noban@127.0.0.1/10"], "Cannot resolve -whitebind address", match=ErrorMatch.PARTIAL_REGEX) def check_tx_relay(self): block_op_true = self.nodes[0].getblock(self.nodes[0].generatetoaddress(100, ADDRESS_BCRT1_P2WSH_OP_TRUE)[0]) self.sync_all() self.log.debug("Create a connection from a whitelisted wallet that rebroadcasts raw txs") # A python mininode is needed to send the raw transaction directly. If a full node was used, it could only # rebroadcast via the inv-getdata mechanism. However, even for whitelisted connections, a full node would # currently not request a txid that is already in the mempool. self.restart_node(1, extra_args=["-whitelist=forcerelay@127.0.0.1"]) p2p_rebroadcast_wallet = self.nodes[1].add_p2p_connection(P2PDataStore()) self.log.debug("Send a tx from the wallet initially") tx = FromHex( CTransaction(), self.nodes[0].createrawtransaction( inputs=[{ 'txid': block_op_true['tx'][0], 'vout': 0, }], outputs=[{ ADDRESS_BCRT1_P2WSH_OP_TRUE: 5, }]), ) tx.wit.vtxinwit = [CTxInWitness()] tx.wit.vtxinwit[0].scriptWitness.stack = [CScript([OP_TRUE])] txid = tx.rehash() self.log.debug("Wait until tx is in node[1]'s mempool") p2p_rebroadcast_wallet.send_txs_and_test([tx], self.nodes[1]) self.log.debug("Check that node[1] will send the tx to node[0] even though it is already in the mempool") connect_nodes(self.nodes[1], 0) with self.nodes[1].assert_debug_log(["Force relaying tx {} from whitelisted peer=0".format(txid)]): p2p_rebroadcast_wallet.send_txs_and_test([tx], self.nodes[1]) wait_until(lambda: txid in self.nodes[0].getrawmempool()) self.log.debug("Check that node[1] will not send an invalid tx to node[0]") tx.vout[0].nValue += 1 txid = tx.rehash() p2p_rebroadcast_wallet.send_txs_and_test( [tx], self.nodes[1], success=False, reject_reason='Not relaying non-mempool transaction {} from whitelisted peer=0'.format(txid), ) def checkpermission(self, args, expectedPermissions, whitelisted): self.restart_node(1, args) connect_nodes(self.nodes[0], 1) peerinfo = self.nodes[1].getpeerinfo()[0] assert_equal(peerinfo['whitelisted'], whitelisted) assert_equal(len(expectedPermissions), len(peerinfo['permissions'])) for p in expectedPermissions: if not p in peerinfo['permissions']: raise AssertionError("Expected permissions %r is not granted." % p) def replaceinconfig(self, nodeid, old, new): with open(self.nodes[nodeid].umkoinconf, encoding="utf8") as f: newText = f.read().replace(old, new) with open(self.nodes[nodeid].umkoinconf, 'w', encoding="utf8") as f: f.write(newText) if __name__ == '__main__': P2PPermissionsTests().main()
[ "vmta@yahoo.com" ]
vmta@yahoo.com
6e69cbf407d691845da736d0f0e4fe9b57e0019a
a65c77b44164b2c69dfe4bfa2772d18ae8e0cce2
/test/testgen.py
045d1ec03e086452f960b6967c1fd274939e737d
[]
no_license
dl8sd11/online-judge
553422b55080e49e6bd9b38834ccf1076fb95395
5ef8e3c5390e54381683f62f88d03629e1355d1d
refs/heads/master
2021-12-22T15:13:34.279988
2021-12-13T06:45:49
2021-12-13T06:45:49
111,268,306
1
6
null
null
null
null
UTF-8
Python
false
false
756
py
#!/usr/bin/env python3 from random import randint, shuffle import string import random def randomString(stringLength=10): """Generate a random string of fixed length """ letters = string.ascii_lowercase return ''.join(random.choice(letters) for i in range(stringLength)) def randomTree(sz): edg = [] for i in range(1, sz): edg.append((i+1, randint(0, i-1)+1)) return edg def get_perm(sz): a = [x for x in range(1,sz+1)] shuffle(a) return a def genInt (): return randint(1,10) def getChar (): x = "AKQJ" if randint(1,13) <= 4: return random.choice(x) else : return str(randint(2, 10)) n = randint(3, 4) a = [str(randint(1, 10)) for x in range(n)] print(n) print(" ".join(a))
[ "tmd910607@gmail.com" ]
tmd910607@gmail.com
d2ec3022630495244009b0dfaad6ae6fb279abda
37dcb5cbe43eb2fd1ffe716e6bcf318dce10a275
/RLE/R8C.py
bb7bfec73a62d8bfdc2d2f2ab76e00969219e847
[ "MIT" ]
permissive
nesdoug/SNES_00
488f405d673927443dbbec05cfa5bfae6667d8c0
c5755eb3e8d01c017b962a8ad2b5ec1b7067f18a
refs/heads/master
2021-11-21T10:50:15.410248
2021-11-06T13:59:24
2021-11-06T13:59:24
237,279,311
12
0
null
null
null
null
UTF-8
Python
false
false
9,049
py
#!/usr/bin/python3 # 8 bit RLE compressor # written by Doug Fraker 2020 # for SNES background maps (and other things) # all 8 bit units, and 8-16 bit headers # non-planar and planar combined version # (planar = split all even and odd bytes) # it tests both and outputs only the smaller # eof byte (last of file) of F0 indicates # non-planar. # eof byte FF = planar, which is what we # expect for SNES maps. # one byte header ---- # MM CCCCCC # M - mode, C - count (+1) # 0 - literal, C+1 values (1-64) # 1 - rle run, C+1 times (1-64) # 2 - rle run, add 1 each pass, C+1 times (1-64) # 3 - extend the value count to 2 bytes # 00 lit, 40 rle, 80 plus, F0 special # two byte header ---- # 11 MM CCCC (high) CCCCCCCC (low) # M - mode (as above), C - count (+1) # count 1-4096 # c0 lit big, d0 = rle big, e0 = plus big # F0 - end of data, non-planar # FF - end of data, planar # input binary up to 32768 bytes # note, planar expects an even # of bytes, # and will pad a zero 00 at the end # of an odd number of input bytes. import sys import os def try_rle(out_array): global index global filesize global count global index2 oldindex = index count = 0 byte1 = 0 byte2 = 0 byte3 = 0 while(index < filesize): if(count >= 4095): break if(in_array[index-1] == in_array[index]): count = count + 1 index = index + 1 else: break if (count > 0): # zero is better, leaving it. #output to the out array if(count > 31): # 2 byte header d0 00 byte1 = ((count >> 8) & 0x0f) + 0xd0 byte2 = count & 0xff byte3 = in_array[index-1] out_array[index2] = byte1 index2 = index2 + 1 out_array[index2] = byte2 index2 = index2 + 1 out_array[index2] = byte3 index2 = index2 + 1 else: # 1 byte header 40 byte1 = (count & 0x3f) + 0x40 byte2 = in_array[index-1] out_array[index2] = byte1 index2 = index2 + 1 out_array[index2] = byte2 index2 = index2 + 1 index = index + 1 else: count = 0 index = oldindex def try_plus(out_array): global index global filesize global count global index2 oldindex = index count = 0 start_value = in_array[index-1] byte1 = 0 byte2 = 0 byte3 = 0 while(index < filesize): if(count >= 255): # in the 8 bit version 4095 doesn't make sense break if(in_array[index-1] == in_array[index] - 1): #what about wrap around ? count = count + 1 index = index + 1 else: break if (count > 0): # zero is better, leaving it. #output to the out array if(count > 31): # 2 byte header e0 00 byte1 = ((count >> 8) & 0x0f) + 0xe0 byte2 = count & 0xff byte3 = start_value out_array[index2] = byte1 index2 = index2 + 1 out_array[index2] = byte2 index2 = index2 + 1 out_array[index2] = byte3 index2 = index2 + 1 else: # 1 byte header 80 byte1 = (count & 0x3f) + 0x80 byte2 = start_value out_array[index2] = byte1 index2 = index2 + 1 out_array[index2] = byte2 index2 = index2 + 1 index = index + 1 else: count = 0 index = oldindex def do_literal(out_array): global index global filesize global count global index2 byte1 = 0 byte2 = 0 byte3 = 0 start_index = index-1 count = 0 index = index + 1 while(index < filesize): if(count >= 4094): # 2 less to fix possible error break if((in_array[index-2] == in_array[index-1]) and (in_array[index-1] == in_array[index])): break if(((in_array[index-2] == in_array[index-1] - 1)) and (in_array[index-1] == in_array[index] - 1)): break count = count + 1 index = index + 1 # back up 1, found a repeat, or repeat + 1 count = count - 1 index = index - 1 nearend = filesize - index if (nearend < 2): #end of file, dump rest if (nearend == 1): count = count + 1 index = index + 1 count = count + 1 index = index + 1 if (count >= 0): #output to the out array count2 = count + 1 if(count > 31): # 2 byte header c0 00 byte1 = ((count >> 8) & 0x0f) + 0xc0 byte2 = count & 0xff out_array[index2] = byte1 index2 = index2 + 1 out_array[index2] = byte2 index2 = index2 + 1 for i in range (0,count2): byte3 = in_array[start_index] out_array[index2] = byte3 index2 = index2 + 1 start_index = start_index + 1 else: # 1 byte header 00 byte1 = (count & 0x3f) out_array[index2] = byte1 index2 = index2 + 1 for i in range (0,count2): byte2 = in_array[start_index] out_array[index2] = byte2 index2 = index2 + 1 start_index = start_index + 1 filename = sys.argv[1] newname = filename[0:-4] + ".rle" oldfile = open(filename, 'rb') newfile = open(newname, 'wb') # warning, this may overwrite old file ! filesize = os.path.getsize(filename) print("input filesize = " + str(filesize)) if(filesize > 32768): exit("error, too large. File should be <= 32768 bytes.") if(filesize < 3): exit("error, file too small.") in_array = [0] * 32768 in_array_P = [0] * 32768 out_array_nonP = [0] * 33000 # a little extra, just in case out_array_P = [0] * 33000 #copy to array for i in range (0, filesize): in_array[i] = ord(oldfile.read(1)) # first try non-planar... index = 1 # start at 1, subtract index2 = 0 non_pl_size = 0 count = 0 #main while(index < filesize): count = 0 try_rle(out_array_nonP) # returns with count > 1 if successful if(count == 0): try_plus(out_array_nonP) # returns with count > 1 if successful if(count == 0): do_literal(out_array_nonP) # do final literal, last byte if(index == filesize): #we need 1 more literal out_array_nonP[index2] = 0 index2 = index2 + 1 byte1 = in_array[filesize-1] out_array_nonP[index2] = byte1 index2 = index2 + 1 # put a final f0 - non-planar symbol out_array_nonP[index2] = 0xf0 index2 = index2 + 1 non_pl_size = index2 # try again with planar... filesize_half = (filesize + 1) // 2 # round up, divide by 2 filesize = filesize_half * 2 split_array = [0] * 16384 split_array2 = [0] * 16384 # split the array for i in range (0, filesize_half): j = i * 2 k = j + 1 split_array[i] = in_array[j] # even bytes split_array2[i] = in_array[k] # odd bytes # copy them back # (so I don't have to change the rest of the code) for i in range (0, filesize_half): in_array_P[i] = split_array[i] j = i + filesize_half in_array_P[j] = split_array2[i] # copy out to another array # so I don't have to refactor the original code. #for i in range(0, index2): # out_array_nonP[i] = out_array[i] #copy planar to original for i in range (0, filesize): in_array[i] = in_array_P[i] #reset and rerun planar count = 0 index2 = 0 index = 1 #main again, planar while(index < filesize): count = 0 try_rle(out_array_P) # returns with count > 1 if successful if(count == 0): try_plus(out_array_P) # returns with count > 1 if successful if(count == 0): do_literal(out_array_P) # do final literal, last byte if(index == filesize): #we need 1 more literal out_array_P[index2] = 0 index2 = index2 + 1 byte1 = in_array[filesize-1] out_array_P[index2] = byte1 index2 = index2 + 1 # put a final ff - planar symbol out_array_P[index2] = 0xff # note out_array_P[] is the Planar version # non-planar is out_array_nonP[] index2 = index2 + 1 pl_size = index2 print("planar out size = " + str(pl_size)) print("non-planar out size = " + str(non_pl_size)) a = 0 if(non_pl_size <= pl_size): #3 is smaller, non-planar print("using non-planar...") for i in range (0, non_pl_size): byte1 = out_array_nonP[i] newfile.write(bytes([byte1])) a = non_pl_size else: print("using planar...") for i in range (0, pl_size): byte1 = out_array_P[i] newfile.write(bytes([byte1])) a = pl_size # output percent of original the output is. b = 100.0 * a / filesize b = round(b, 2) print(" new filesize = " + str(a)) print(" compared to orig = " + str(b) + "%") # close the files. oldfile.close newfile.close
[ "dougfraker@gmail.com" ]
dougfraker@gmail.com
e6a758972e70604e79582b0290a099ed73535f7d
0f880611b30941662cee4b37fb16e90b1227a3e4
/lib/utils/log.py
1a50c636269ed470cc8a80b1712004bcd313ad75
[]
no_license
ColdHumour/PortfolioMonitor
6cefa499aa6bec059e035fca7300926f84c840e9
d7da8a8c52e494bee30b1238521fc383b2489154
refs/heads/master
2020-04-06T05:13:51.245235
2016-11-02T07:01:34
2016-11-02T07:01:34
54,858,149
0
1
null
2016-08-24T07:08:16
2016-03-28T01:59:22
Python
UTF-8
Python
false
false
731
py
# -*- coding: utf-8 -*- """ log.py @author: yudi.wu """ import json import logging from . path import LOG_FILE, CONFIG_FILE open(LOG_FILE, 'w').close() def set_logger(name, level): logger = logging.Logger(name) logger.setLevel(level) # set handler handler = logging.FileHandler(LOG_FILE) handler.setLevel(level) # set formatter formatter = logging.Formatter('%(asctime)s [%(levelname)s] File:%(filename)s Line:%(lineno)d - %(message)s') handler.setFormatter(formatter) logger.addHandler(handler) return logger with open(CONFIG_FILE, 'r') as config_file: config = json.load(config_file) loglevel = config["loglevel"] logger = set_logger("APIServer", getattr(logging, loglevel))
[ "jasper.wuyd@gmail.com" ]
jasper.wuyd@gmail.com
e0164f076b0f134294981ab8b376febecefb227c
02f2e8b08231290656ffaea8fce5bf03b2150780
/mall/apps/areas/serializers.py
e74483ac5c6f3463545b6b8434d75d72ba22fb25
[ "MIT" ]
permissive
googleliyang/django-meiduo-teach
407aa08b65d29c914281a8b759db3b1cdd25defb
afb89767b77118f38dca6f1c2e3b98db0474e48f
refs/heads/master
2020-04-28T16:25:23.905191
2019-03-11T04:38:01
2019-03-11T04:38:01
175,409,991
0
0
null
null
null
null
UTF-8
Python
false
false
792
py
from rest_framework import serializers from areas.models import Area # area # 省的信息 class AreaSerializer(serializers.ModelSerializer): # area_set = serializers.PrimaryKeyRelatedField(many=True,read_only=True) class Meta: model = Area fields = ['id','name'] # 市的序列化器 class SubsAreaSerializer(serializers.ModelSerializer): # 想 通过id 来获取这个id所对应的所有的值 area # 同时 转换为 字典 # area_set = serializers.PrimaryKeyRelatedField(many=True,read_only=True) # [1,2,3,4,5] area_set = AreaSerializer(many=True,read_only=True) # subs = AreaSerializer(many=True,read_only=True) class Meta: model = Area # fields = ['subs','id','name'] fields = ['area_set','id','name']
[ "qiruihua@itcast.cn" ]
qiruihua@itcast.cn
414e11c2131f405f7fa99aba0b23eed0f5667a6f
e6a43cb50cd36ed8bc3f70a3cd830bffb3f03d5e
/set03/q23_sectionlv.py
4b077194b48d1e5b797178b986dcb9b80fcb16aa
[]
no_license
pizzaboi/nlp100_2_python
a74b4ab0aadffe6a5cd75cd205bca8777ebce8e5
b36a50442b581c6547ab6f02c7c2b18038a7b420
refs/heads/master
2016-09-05T09:31:40.081775
2015-12-17T05:53:29
2015-12-17T05:53:29
34,831,332
0
0
null
null
null
null
UTF-8
Python
false
false
321
py
#! /usr/bin/python #-*- coding: utf-8 -*- """ USAGE: python q20_readjson.py < jawiki-country.json.gz | python q23_sectionlv.py """ import re import sys def section_lv(): for line in sys.stdin: if line.startswith('=='): print line.strip('= \n'), (line.count('=') / 2) - 1 if __name__ == '__main__': section_lv()
[ "tophamhatt.crazy@gmail.com" ]
tophamhatt.crazy@gmail.com
80c5331bac48eab4818a62eed9690372b32be540
0fbc1c6f72a69f368c14f2ca9af8b03487140b6d
/contact/views.py
a057a95c4ae29a16f497435a13015110b9c9b17d
[]
no_license
ritusabu/Ecommerce-product
1c3e95d09438ab2ff8dfb83aaf3c2771f97e3440
b9e6f8f16c1dbf19a0f2624f5c9b1bbdd8b2d670
refs/heads/master
2023-03-05T05:18:02.065448
2021-02-16T10:30:10
2021-02-16T10:30:10
334,204,268
1
0
null
null
null
null
UTF-8
Python
false
false
664
py
from django.shortcuts import render from django.http import HttpResponse from contact.models import Contact from product.models import Offer, Category, Products # Create your views here. def contact(request): a= request.POST.get("name", "") b= request.POST.get("mail", "") c= request.POST.get("text", "") o= Offer.objects.all() t= Category.objects.all() if a!="" and b!="" and c!="": c=Contact(name=a, email=b, messege=c) c.save() context={ "offer":o, "catagory":t } return render(request, 'contact/contact.html' , context) def home(request): return HttpResponse("E-commerse")
[ "ritusabu12@gmail.com" ]
ritusabu12@gmail.com
fff66fda560450668aa9428a18225497f0cf09ee
cbb6ef5b109d262459167dc8cee8264e4b4b4487
/pollsapi/polls/migrations/0001_initial.py
e72942e34d8f5ed5363e047f78baa698c959b1a4
[]
no_license
TaTRaTa/Django_DRF
2b5541eec39fbc85fa71634f80efadf80ff15a36
232d36058073ee74e63e13cf9f69481b51d0a3b9
refs/heads/master
2021-03-13T12:38:16.164786
2020-03-17T21:56:32
2020-03-17T21:56:32
246,681,697
0
0
null
null
null
null
UTF-8
Python
false
false
1,910
py
# Generated by Django 3.0.4 on 2020-03-12 22:08 from django.conf import settings from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): initial = True dependencies = [ migrations.swappable_dependency(settings.AUTH_USER_MODEL), ] operations = [ migrations.CreateModel( name='Poll', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('question', models.CharField(max_length=100)), ('pub_date', models.DateTimeField(auto_now=True)), ('created_by', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], ), migrations.CreateModel( name='Choice', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('choice_text', models.CharField(max_length=100)), ('poll', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='choices', to='polls.Poll')), ], ), migrations.CreateModel( name='Vote', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('choice', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='votes', to='polls.Choice')), ('poll', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='polls.Poll')), ('voted_by', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to=settings.AUTH_USER_MODEL)), ], options={ 'unique_together': {('poll', 'voted_by')}, }, ), ]
[ "cvetomir.defyy@gmail.com" ]
cvetomir.defyy@gmail.com
2c4d4ae9918a8cfb4516783ec1911a72d0c2c3c0
ca51878f71e5b3e1958e42e45019f92c9045b02e
/ClubWebsite/settings.py
299f6528cb142d270cc7732323a0fb9ce4ec8ea1
[ "Apache-2.0" ]
permissive
pettta/ClubWebsite
07d1c828fa73f8d5b35fba99bfa910ba3a610b7c
83ed5e2c923d8c13aa1f76e564bb817e6ad81fb1
refs/heads/main
2023-07-16T10:43:55.841813
2021-07-01T20:32:46
2021-07-01T20:32:46
401,377,529
0
0
Apache-2.0
2021-08-30T14:41:02
2021-08-30T14:41:01
null
UTF-8
Python
false
false
3,280
py
""" Django settings for ClubWebsite project. Generated by 'django-admin startproject' using Django 3.2.5. For more information on this file, see https://docs.djangoproject.com/en/3.2/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/3.2/ref/settings/ """ from pathlib import Path import os # Build paths inside the project like this: BASE_DIR / 'subdir'. BASE_DIR = Path(__file__).resolve().parent.parent # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/3.2/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'django-insecure-g7$1p+p7425l%g=$v-g22csj)wn3&e7s*7!#3=cd9_xswsm061' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True ALLOWED_HOSTS = ['*'] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', ] MIDDLEWARE = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'ClubWebsite.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', ], }, }, ] WSGI_APPLICATION = 'ClubWebsite.wsgi.application' # Database # https://docs.djangoproject.com/en/3.2/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } # Password validation # https://docs.djangoproject.com/en/3.2/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/3.2/topics/i18n/ LANGUAGE_CODE = 'en-us' TIME_ZONE = 'UTC' USE_I18N = True USE_L10N = True USE_TZ = True # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/3.2/howto/static-files/ STATIC_URL = '/static/' # Default primary key field type # https://docs.djangoproject.com/en/3.2/ref/settings/#default-auto-field DEFAULT_AUTO_FIELD = 'django.db.models.BigAutoField'
[ "vivekkhimani07@gmail.com" ]
vivekkhimani07@gmail.com
efcef61de6ca47fda0549982e2032134ed07f440
f9be26f1a01aecda9f4ded28922e1490c450558d
/plc-2.0/lib/python2.7/site-packages/pyfits-3.5-py2.7-linux-x86_64.egg/pyfits/compression.py
8b72aebbd02c0324116a1c1f29b8765231a7a106
[]
no_license
NinaKate/CCR
7c71c5dc2a31af2bb76b2cc2d9e5a7c6193486ee
ba4c5ae159d97bcc78eb3570428ec33f87761ebe
refs/heads/master
2018-11-14T01:24:27.890646
2018-08-27T17:08:23
2018-08-27T17:08:23
105,586,700
0
0
null
2017-12-03T20:20:34
2017-10-02T21:24:35
C++
UTF-8
Python
false
false
281
py
def __bootstrap__(): global __bootstrap__, __loader__, __file__ import sys, pkg_resources, imp __file__ = pkg_resources.resource_filename(__name__,'compression.so') __loader__ = None; del __bootstrap__, __loader__ imp.load_dynamic(__name__,__file__) __bootstrap__()
[ "ninastei@buffalo.edu" ]
ninastei@buffalo.edu
21b6f5c9313bb2beb0284929069e54b107bf6ea9
1d7a6d74a9a906a96aa61ab94c1edee610ccd92b
/zip_build/rvt-qgis/qrvt_dialog.py
36b723c750e4f016c2627a61d72d8aac9017b43b
[ "Apache-2.0", "LicenseRef-scancode-unknown-license-reference" ]
permissive
IsaacMrSmile/rvt-qgis
350fc6283cfe17e021a22ec62ec98e000b2f0606
fac15ef95534483967be691a1dfacd5b0bd9a047
refs/heads/master
2023-02-21T00:32:30.917054
2021-01-22T07:12:22
2021-01-22T07:12:22
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,105
py
# -*- coding: utf-8 -*- """ /*************************************************************************** QRVTDialog A QGIS plugin RVT plugin lets you compute different visualizations from raster DEM. Generated by Plugin Builder: http://g-sherman.github.io/Qgis-Plugin-Builder/ ------------------- begin : 2020-10-12 git sha : $Format:%H$ copyright : (C) 2020 by Research Centre of the Slovenian Academy of Sciences and Arts email : ziga.kokalj@zrc-sazu.si ***************************************************************************/ /*************************************************************************** * * * This program is free software; you can redistribute it and/or modify * * it under the terms of the GNU General Public License as published by * * the Free Software Foundation; either version 2 of the License, or * * (at your option) any later version. * * * ***************************************************************************/ """ import os from qgis.PyQt import uic from qgis.PyQt import QtWidgets # This loads your .ui file so that PyQt can populate your plugin with the elements from Qt Designer FORM_CLASS, _ = uic.loadUiType(os.path.join( os.path.dirname(__file__), 'qrvt_dialog_base.ui')) class QRVTDialog(QtWidgets.QDialog, FORM_CLASS): def __init__(self, parent=None): """Constructor.""" super(QRVTDialog, self).__init__(parent) # Set up the user interface from Designer through FORM_CLASS. # After self.setupUi() you can access any designer object by doing # self.<objectname>, and you can use autoconnect slots - see # http://qt-project.org/doc/qt-4.8/designer-using-a-ui-file.html # #widgets-and-dialogs-with-auto-connect self.setupUi(self)
[ "zm8597@student.uni-lj.si" ]
zm8597@student.uni-lj.si
5d03c3f6f21cf2a5cf29fc8907a7adfcc620d57f
2ad41c2a31618433568c86e63f68a3ef2918d55c
/tool/Modules/cfg_scripts.py
25ca07351b013433ffe1409fb953f7919d31d99b
[ "MIT" ]
permissive
Iemnur/megaman-zx-traducao-ptbr
7cad0b7f7bcfd6692fe850f3c6c4e26ab2b90f63
f2710a06052384cf93d423681e9875c6cd424f06
refs/heads/master
2021-12-14T20:13:48.206022
2020-05-26T01:53:10
2020-05-26T01:53:10
82,298,127
0
0
null
null
null
null
UTF-8
Python
false
false
598
py
''' Created on 05/03/2013 @author: diego.hahn ''' import os.path import sys python_path = os.path.dirname( sys.executable ) packages_path = os.path.join( python_path , r"Lib\site-packages" ) scripts_path = os.path.dirname( os.path.abspath( __file__ ) ) libs = [r"" , r"rhCompression", r"rhImages", r"pytable"] with open( os.path.join( packages_path , "mylibs.pth" ), "w" ) as pth: for lib in libs: lib_path = os.path.join( scripts_path, lib ) if os.path.isdir( lib_path ): print( ">>> Adding %s to pth file" % lib ) pth.write( "%s\n" % lib_path )
[ "hansen.hahn@gmail.com" ]
hansen.hahn@gmail.com
ac572ec20daedd8a6e8d42a53fc9d042e52f7602
24820f2df644263d6273e7de78f0733ca81511d9
/Robot_Simulator/pose_script.py
d28af4ff074b8220be03e92cd55a9d617ebc6258
[]
no_license
dwhisler/robotic_arm
1c0603d9a21710bd08d108b3397aa7147f789720
c8eef383e1401141f5a87f485798240cfd2ff8fb
refs/heads/master
2020-03-15T18:37:24.885097
2018-09-20T04:03:42
2018-09-20T04:03:42
132,288,033
0
0
null
null
null
null
UTF-8
Python
false
false
836
py
# -*- coding: utf-8 -*- """ Created on Wed Sep 19 23:18:15 2018 @author: David """ from ServoDriver import * from DTRobot import * COM = 4 serv = ServoDriver(COM) rob = DTRobot('config.ini') #serv.setPose([0, 0, 0, 0, 0, 0]) #input("Press Enter to continue...") serv.setPose(rob.robot.joints) input("Press Enter to continue...") while(1): currentPose = rob.robot.fk(rob.robot.joints) trans = translation_matrix([0,-0.1,0]) newPose = np.matmul(trans,currentPose) print(currentPose) print(trans) print(newPose) input("Press Enter to continue...") newJointAngles = rob.robot.ik(newPose) print(newJointAngles) input("Press Enter to continue...") serv.setPose(newJointAngles) rob.robot.joints = newJointAngles input("Press Enter to continue...") serv.ser.close()
[ "dwhis428@gmail.com" ]
dwhis428@gmail.com
f7fcb553c02ffff0e4816ffbb847e1c926470726
b55f70755712b26688b80a8ba3806a4124fbcd11
/BinaryTree/lowest_common_ancestor.py
c5fac7a034bae171afb4a6a2bb03b6ce00e81aa2
[]
no_license
Shanshan-IC/Algorithm_Python
a44703a0f33370c47e3e55af70aadeae08d5a1a5
ace23976d2f1f51141498c4c4ea6bca0039b233f
refs/heads/master
2021-09-08T07:16:59.576674
2018-03-08T09:24:01
2018-03-08T09:24:01
114,254,497
0
0
null
null
null
null
UTF-8
Python
false
false
926
py
''' 两个值都在左边,则LCA在左边 两个值都在右边,则LCA在右边 一个在左一个在右,则说明LCA就是当前的root节点。 ''' """ Definition of TreeNode: class TreeNode: def __init__(self, val): self.val = val self.left, self.right = None, None """ class Solution: """ @param: root: The root of the binary search tree. @param: A: A TreeNode in a Binary. @param: B: A TreeNode in a Binary. @return: Return the least common ancestor(LCA) of the two nodes. """ def lowestCommonAncestor(self, root, A, B): if not root or root is A or root is B: return root left = self.lowestCommonAncestor(root.left, A, B) right = self.lowestCommonAncestor(root.right, A, B) if left and right: return root if left: return left if right: return right return None
[ "shanshan.fu15@imperial.ac.uk" ]
shanshan.fu15@imperial.ac.uk
f8041e3723b6c11795df50a6b0361b01f3581dbe
c81a507a4c76db54e9e29a2a457a017a8725a8c2
/scripts/data/CIFAR10.py
1b8661fae5a1a2e9020944f726264ac045937008
[]
no_license
CSWater/blitz
b8e5d1f5a69a64a9dc12be7252f95ed40f2178c5
cc5488f1623f5b3161fa334e6813d499918dcc5e
refs/heads/master
2020-06-11T00:27:52.466242
2016-12-27T07:49:22
2016-12-27T07:49:22
75,832,706
1
0
null
2016-12-07T12:11:00
2016-12-07T12:10:59
null
UTF-8
Python
false
false
3,580
py
import os import sys import cPickle import h5py import numpy as np def _valid_path_append(path, *args): full_path = os.path.expanduser(path) res = [] if not os.path.exists(full_path): os.makedirs(full_path) if not os.path.isdir(full_path): raise ValueError("path: {0} is not a valid directory".format(path)) for suffix_path in args: res.append(os.path.join(full_path, suffix_path)) if len(res) == 0: return path elif len(res) == 1: return res[0] else: return res def load_cifar10(path="./data", normalize=True, contrast_normalize=False, whiten=False): cifar = dataset_meta['cifar-10'] workdir, filepath = _valid_path_append(path, '', cifar['file']) batchdir = os.path.join(workdir, '') train_batches = [os.path.join(batchdir, 'data_batch_' + str(i)) for i in range(1, 6)] Xlist, ylist = [], [] for batch in train_batches: with open(batch, 'rb') as f: d = cPickle.load(f) Xlist.append(d['data']) ylist.append(d['labels']) X_train = np.vstack(Xlist) y_train = np.vstack(ylist) with open(os.path.join(batchdir, 'test_batch'), 'rb') as f: d = cPickle.load(f) X_test, y_test = d['data'], d['labels'] y_train = y_train.reshape(-1, 1) y_test = np.array(y_test).reshape(-1, 1) num_train = y_train.shape[0] num_test = y_test.shape[0] y_train_new = np.zeros((num_train, 10)) y_test_new = np.zeros((num_test, 10)) for col in range(10): y_train_new[:, col] = y_train[:,0] = col y_test_new[:, col] = y_test[:,0] = col if contrast_normalize: norm_scale = 55.0 X_train = global_contrast_normalize(X_train, scale=norm_scale) X_test = global_contrast_normalize(X_test, scale=norm_scale) if normalize: X_train = X_train / 255. X_test = X_test / 255. if whiten: zca_cache = os.path.join(workdir, 'cifar-10-zca-cache.pkl') X_train, X_test = zca_whiten(X_train, X_test, cache=zca_cache) #save the hdf5 files repo_path = os.path.expandvars(os.path.expanduser(workdir)) save_dir = os.path.join(repo_path, 'HDF5') if not os.path.exists(save_dir): os.makedirs(save_dir) fname = os.path.join(save_dir, 'train_data.h5') file_train_data = h5py.File(fname, 'w') fname = os.path.join(save_dir, 'train_label.h5') file_train_label = h5py.File(fname, 'w') fname = os.path.join(save_dir, 'test_data.h5') file_test_data = h5py.File(fname, 'w') fname = os.path.join(save_dir, 'test_label.h5') file_test_label = h5py.File(fname, 'w') file_train_data.create_dataset('data', data = X_train) file_train_data.create_dataset('sample_num', data = num_train) file_train_label.create_dataset('data', data = y_train_new) file_train_label.create_dataset('sample_num', data = num_train) file_test_data.create_dataset('data', data = X_test) file_test_data.create_dataset('sample_num', data = num_test) file_test_label.create_dataset('data', data = y_test_new) file_test_label.create_dataset('sample_num', data = num_test) file_train_data.close() file_train_label.close() file_test_data.close() file_test_label.close() return (X_train, y_train_new), (X_test, y_test_new), 10 dataset_meta = { 'cifar-10': { 'size': 170498071, 'file': 'cifar-10-python.tar.gz', 'url': 'http://www.cs.toronto.edu/~kriz', 'func': load_cifar10 } } if __name__ == '__main__': load_cifar10()
[ "robinho364@gmail.com" ]
robinho364@gmail.com
be2c9c0203597021595d503e7f6d72164b2028eb
889b542e025a3f4857aa9faf7abcec6afebdce37
/init.spec
c2ada7cfd4de5d92f4af1c8ed130636b9c6554cb
[ "MIT" ]
permissive
mfneirae/Sara
c685bafe6bd69378cd81ada98d011d7f18564948
a4b4bcae05bf39e4f8ad49b47640ccf552abce03
refs/heads/master
2020-03-25T19:20:04.915517
2018-09-19T21:43:24
2018-09-19T21:43:24
144,077,388
0
0
null
null
null
null
UTF-8
Python
false
false
813
spec
# -*- mode: python -*- block_cipher = None a = Analysis(['init.py'], pathex=['C:\\Users\\Ing_99\\Desktop\\Sara'], binaries=[], datas=[], hiddenimports=[], hookspath=[], runtime_hooks=[], excludes=[], win_no_prefer_redirects=False, win_private_assemblies=False, cipher=block_cipher, noarchive=False) pyz = PYZ(a.pure, a.zipped_data, cipher=block_cipher) exe = EXE(pyz, a.scripts, a.binaries, a.zipfiles, a.datas, [], name='init', debug=False, bootloader_ignore_signals=False, strip=False, upx=True, runtime_tmpdir=None, console=True )
[ "mfneirae@unal.edu.co" ]
mfneirae@unal.edu.co
7fd98a807bdf12bc76653e933673247354cf7d49
692a43899841bf0569c4f44bf2bba435eea71bbe
/pycloud/pycloud/network/cloudlet_dns.py
3850491cebff1b2b95defae29992e92f2166ad82
[ "LicenseRef-scancode-warranty-disclaimer", "MIT", "BSD-2-Clause" ]
permissive
SEI-TAS/pycloud
52a4403cdbdb2eaeabf18c377a806089591aea31
beebe297b3ed35de0bc53b75f9f2b4a2e80d892a
refs/heads/master
2023-06-22T08:19:46.038367
2023-06-15T18:24:31
2023-06-15T18:24:31
30,374,770
5
4
NOASSERTION
2023-06-15T18:24:32
2015-02-05T19:46:50
Python
UTF-8
Python
false
false
3,913
py
# KVM-based Discoverable Cloudlet (KD-Cloudlet) # Copyright (c) 2015 Carnegie Mellon University. # All Rights Reserved. # # THIS SOFTWARE IS PROVIDED "AS IS," WITH NO WARRANTIES WHATSOEVER. CARNEGIE MELLON UNIVERSITY EXPRESSLY DISCLAIMS TO THE FULLEST EXTENT PERMITTEDBY LAW ALL EXPRESS, IMPLIED, AND STATUTORY WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, AND NON-INFRINGEMENT OF PROPRIETARY RIGHTS. # # Released under a modified BSD license, please see license.txt for full terms. # DM-0002138 # # KD-Cloudlet includes and/or makes use of the following Third-Party Software subject to their own licenses: # MiniMongo # Copyright (c) 2010-2014, Steve Lacy # All rights reserved. Released under BSD license. # https://github.com/MiniMongo/minimongo/blob/master/LICENSE # # Bootstrap # Copyright (c) 2011-2015 Twitter, Inc. # Released under the MIT License # https://github.com/twbs/bootstrap/blob/master/LICENSE # # jQuery JavaScript Library v1.11.0 # http://jquery.com/ # Includes Sizzle.js # http://sizzlejs.com/ # Copyright 2005, 2014 jQuery Foundation, Inc. and other contributors # Released under the MIT license # http://jquery.org/license import dynamic_dns import os from pycloud.pycloud.network.tsig import load_tsig_key SVMS_ZONE_NAME = 'svm.cloudlet.local.' CLOUDLET_HOST_NAME = 'cloudlet' # Internal file path, relative to data folder. KEY_FILE_PATH = 'dns/Ksvm.cloudlet.local.private' ################################################################################################################# # Object used to manage the cloudlet DNS server. ################################################################################################################# class CloudletDNS(object): ################################################################################################################# # Constructor. ################################################################################################################# def __init__(self, root_data_folder): full_path = os.path.join(os.path.abspath(root_data_folder), KEY_FILE_PATH) self.key = load_tsig_key(full_path) ################################################################################################################# # Generates a FQDN in our SVM zone from a hostname. ################################################################################################################# @staticmethod def generate_fqdn(hostname): return hostname + "." + SVMS_ZONE_NAME ################################################################################################################# # Registers an SVM. ################################################################################################################# def register_svm(self, svm_fqdn, ip_address=None): # Depending on networking mode, we will need to register an explicit A record with an IP, or a cname to cloudlet. if ip_address: record_value = ip_address record_type = 'A' else: record_value = CLOUDLET_HOST_NAME record_type = 'CNAME' if not self.key: print "Can't register SVM: TSIG key not loaded." return dynamic_dns.add_dns_record(SVMS_ZONE_NAME, self.key, svm_fqdn, record_value, record_type=record_type) ################################################################################################################# # Unregisters an SVM. ################################################################################################################# def unregister_svm(self, svm_fqdn): if not self.key: print "Can't unregister SVM: TSIG key not loaded." return dynamic_dns.remove_dns_record(SVMS_ZONE_NAME, self.key, svm_fqdn)
[ "sebastian.echeverria@gmail.com" ]
sebastian.echeverria@gmail.com
d080f213409dd5c85811969f8b2e789bd4d38370
78b165272ba2cb3719de3c27e83d9c5b3b6b0a85
/scripts/python/minheap.py
9e4ddae3174b98e471e0e7a2d94196a4efea0227
[]
no_license
SeaTalk/Funcodes
5fcd4727565d2ca99d226b04bd58385218bd2fbf
bd34b19fa99be42433e1305f2f8c6bef14a6bb3f
refs/heads/master
2021-10-11T22:11:18.898476
2021-09-30T07:26:34
2021-09-30T07:26:34
140,306,910
2
1
null
null
null
null
UTF-8
Python
false
false
2,082
py
#!/usr/bin/python import string class MinHeap(object): def __init__(self, cap): self.data=[] self.count=len(self.data) self.capacity=cap def size(self): return self.count def isEmpty(self): return self.count == 0 def getData(self): return self.data def insert(self, item): if self.count < self.capacity: self.data.append(item) self.count += 1 if self.count == self.capacity: return self.build_heap() return 0 #self.shiftup(self.count) if item[1] > self.data[0][1] : self.data[0] = item return self.fix_heap(0) def build_heap(self): if self.count < 1: return -1 middle = (self.count-1)/2 while middle >= 0 : tmp_index = (middle << 1) + 1 tmp_index2 = (middle + 1) << 1 tmp = middle if tmp_index2 < self.count and self.data[tmp_index][1] > self.data[tmp_index2][1]: tmp = tmp_index2 else : tmp = tmp_index if tmp < self.count and self.data[tmp][1] < self.data[middle][1]: self.data[middle], self.data[tmp] = self.data[tmp], self.data[middle] middle = middle - 1 return 0 def fix_heap(self, ind): index = ind hasChild = True while index < self.count and hasChild == True: j = (index << 1) + 1 k = (index + 1) << 1 tmp = j hasChild = False if k < self.count and self.data[j][1] > self.data[k][1]: tmp = k hasChild = True else : tmp = j hasChild = True if tmp < self.count and self.data[tmp][1] < self.data[index][1]: self.data[index], self.data[tmp] = self.data[tmp], self.data[index] hasChild = True index = tmp return 0
[ "tangjintao@jd.com" ]
tangjintao@jd.com
848df2db66999946ac0072ad26a7281ea8eebf06
7a110d64c3d1ca8a88623b6fb0453904011552ad
/weather_app/weather/views.py
75ca56cb64cdd7fbcaeccffc2ff30c9f06367b9b
[]
no_license
shubhamt10/weather-app-django
a93e15ce146fd8fcbd0aaaed7166eb418d97d244
539a4d087f88d78621f456aec5ef1a6f261ae4f0
refs/heads/master
2021-03-25T11:44:52.885816
2020-03-16T05:09:35
2020-03-16T05:09:35
247,614,554
0
0
null
null
null
null
UTF-8
Python
false
false
1,871
py
from django.shortcuts import render,redirect import requests from .models import City from .forms import CityForm # Create your views here. def index(request): url = 'http://api.openweathermap.org/data/2.5/find?q={}&units=metric&APPID=cc40415657293a697c8f80bd1e548a4b' err_msg = '' message = '' message_class = '' if request.method == 'POST': form = CityForm(request.POST) if form.is_valid(): new_city = form.cleaned_data['name'] existing_city_count = City.objects.filter(name=new_city).count() if existing_city_count == 0: r = requests.get(url.format(new_city)).json() if(len(r['list']) == 0): err_msg = 'City does not exist' else: form.save() else: err_msg = 'City already exists in the database!' if err_msg: message = err_msg message_class = 'is-danger' else: message = 'City added successfully' message_class = 'is-success' form = CityForm() cities = City.objects.all() weather_data = [] for city in cities: r = requests.get(url.format(city)).json() data = r['list'][0] city_weather = { 'city' : data['name'], 'temperature' : data['main']['temp'], 'description' : data['weather'][0]['description'], 'icon' : data['weather'][0]['icon'], } weather_data.append(city_weather) context = { 'weather_data' : weather_data, 'form' : form, 'message' : message, 'message_class' : message_class } return render(request,'weather/weather.html',context) def delete_city(request,city_name): City.objects.get(name=city_name).delete() return redirect('home')
[ "tripathi.shubham509@gmail.com" ]
tripathi.shubham509@gmail.com
fee35bcc21bb856c0b4260204cdaba7d991b37b3
8ead622826bc21b37eb6fd2048af91d510151356
/projects/alonememo/app.py
94e5e9ca5d90d447b61d5b72d5f09acf303d21a0
[]
no_license
smc5720/Sparta-Coding-Basic
5762a5e397c60fcbde5e9ea58fcb9449931d8031
ff4271ededa516ed5c44875509e4b916f04320c5
refs/heads/main
2023-05-14T19:21:53.848721
2021-06-11T09:32:00
2021-06-11T09:32:00
374,037,883
1
0
null
null
null
null
UTF-8
Python
false
false
1,456
py
from flask import Flask, render_template, jsonify, request app = Flask(__name__) import requests from bs4 import BeautifulSoup from pymongo import MongoClient client = MongoClient('localhost', 27017) db = client.dbsparta ## HTML을 주는 부분 @app.route('/') def home(): return render_template('index.html') @app.route('/memo', methods=['GET']) def listing(): articles = list(db.articles.find({}, {'_id': False})) return jsonify({'all_articles': articles}) ## API 역할을 하는 부분 @app.route('/memo', methods=['POST']) def saving(): url_receive = request.form['url_give'] comment_receive = request.form['comment_give'] headers = { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64)AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36'} data = requests.get(url_receive, headers=headers) soup = BeautifulSoup(data.text, 'html.parser') og_title = soup.select_one('meta[property="og:title"]')['content'] og_image = soup.select_one('meta[property="og:image"]')['content'] og_desc = soup.select_one('meta[property="og:description"]')['content'] doc = { 'title': og_title, 'image': og_image, 'desc': og_desc, 'url': url_receive, 'comment': comment_receive, } db.articles.insert_one(doc) return jsonify({'msg': '저장 완료'}) if __name__ == '__main__': app.run('0.0.0.0', port=5000, debug=True)
[ "smc503@naver.com" ]
smc503@naver.com
d8ec9342e39bee550628a862353198ee5b37bcd4
cd2a90eb6df7361a2ed00312ff1be196bf686f0e
/homework/conftest.py
389e88c0dba10d7f481700966d7f3336333c57db
[]
no_license
Liuzicheng1994/auto
8006337b530b80fea39cd36a04bf103e2a3ba711
e9a4069f454d0be0e5be2dd79700944623dea153
refs/heads/master
2023-07-30T17:59:28.558035
2021-09-10T01:40:00
2021-09-10T01:40:00
356,810,977
0
0
null
null
null
null
UTF-8
Python
false
false
180
py
import pytest from homework.calculator import Caculator @pytest.fixture() def calculate(): print("开始计算") cal =Caculator() yield cal print("结束计算")
[ "569974494@qq.com" ]
569974494@qq.com
f50377730a35ff7aa5b58fa06bcf47fcd71189ea
033da72a51c76e5510a06be93229a547a538cf28
/Data Engineer with Python Track/20. Introduction to Spark SQL in Python/Chapter/01. Pyspark SQL/02-Determine the column names of a table.py
a60646c8daa0abfe3fe390558fd3a17b52d8658c
[]
no_license
ikhwan1366/Datacamp
d5dcd40c1bfeb04248977014260936b1fb1d3065
7738614eaebec446842d89177ae2bc30ab0f2551
refs/heads/master
2023-03-06T13:41:06.522721
2021-02-17T22:41:54
2021-02-17T22:41:54
null
0
0
null
null
null
null
UTF-8
Python
false
false
638
py
''' Determine the column names of a table The video lesson showed how to run an SQL query. It also showed how to inspect the column names of a Spark table using SQL. This is important to know because in practice relational tables are typically provided without additional documentation giving the table schema. Don't hesitate to refer to the slides available at the right of the console if you forget how something was done in the video. Instructions 100 XP - Use a DESCRIBE query to determine the names and types of the columns in the table schedule. ''' # Inspect the columns in the table df spark.sql("DESCRIBE schedule").show()
[ "surel.chandrapratama@gmail.com" ]
surel.chandrapratama@gmail.com
0aa0ebe728e983713ebd5eb41983a47e04759129
d1f4ab60c5439c4e91025d96384acdedbcb47271
/DFS,BFS/5-6 미로 탈출.py
8d23a0a4c92327dfb359e8bd664531de2d099f1c
[]
no_license
ans4572/CodingTest-with-Python
8722f07967c76073485282c4eee7506a6d75edb6
ab5e559fd2b8bd73785435c60debd6d558b01ca7
refs/heads/main
2023-07-31T10:24:01.685112
2021-09-24T16:13:34
2021-09-24T16:13:34
372,752,459
0
0
null
null
null
null
UTF-8
Python
false
false
580
py
from collections import deque N,M = map(int,input().split()) graph = [] for i in range(N): graph.append(list(map(int,input()))) dx = [-1,0,1,0] dy = [0,1,0,-1] def BFS(x,y): queue = deque() queue.append((x,y)) while queue: x, y = queue.popleft() for i in range(4): nx = x + dx[i] ny = y + dy[i] if nx >= 0 and nx < N and ny >= 0 and ny < M and graph[nx][ny] == 1: graph[nx][ny] = graph[x][y] + 1 queue.append((nx,ny)) BFS(0,0) print(graph[N-1][M-1])
[ "noreply@github.com" ]
noreply@github.com
69d495704dd2fc3552f5d20c65f2c842be179e23
164ebfa24817cd7d4c5dbf760c1600232aebee1c
/src/filer/observers/observer.py
086f2765aafd9ab0ef6d4cd8cfb8657ebcc293d2
[]
no_license
adarshtri/filer
bd3806e4d8b3859b5cc68a30ed4ee2e46178342d
985fcb9a9792ba18603771489bb6dbdae19d638d
refs/heads/master
2020-05-19T05:49:34.631133
2019-05-04T06:08:05
2019-05-04T06:08:05
184,858,317
0
0
null
null
null
null
UTF-8
Python
false
false
322
py
from interface import Interface class Observer(Interface): """Interface to support observers for newly registered directories""" def update(self, updateinfo: dict): """ :param updateinfo: dictionary containing information regarding the updation. :return: None """ pass
[ "adarsh.trivedi100@gmail.com" ]
adarsh.trivedi100@gmail.com
f8b10ec3dbc9c3b2d9a88a5d5b82dd518963aa1a
155b365fb459caff5f57f9e5eb55a26895a016cd
/evento_01/evento_01/wsgi.py
a69659105175ae9239e21c792c5af9e5efe7552a
[]
no_license
Claison/EVENTO
a43fa39eb3ea4fdd29c4c9ffd858b6e20be1c02b
61b97f980f7cd77b4de92088c6601378a520ea86
refs/heads/master
2021-06-25T18:55:51.229559
2017-09-12T00:59:17
2017-09-12T00:59:17
103,141,973
0
0
null
null
null
null
UTF-8
Python
false
false
396
py
""" WSGI config for evento_01 project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/1.11/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault("DJANGO_SETTINGS_MODULE", "evento_01.settings") application = get_wsgi_application()
[ "30608657+Claison@users.noreply.github.com" ]
30608657+Claison@users.noreply.github.com
229ad85295f1114269334096a485f0e1a78646d3
17d7d9a35c617132965f331060e846d9fe12f129
/lessons/16/serialization2.py
f93ed3958b09fb964dae1bb369120a7ae61b2498
[]
no_license
loristissino/oopython
c46f3e07dc6c4829362efb43f5bbfb496053747b
0e3ab872f8c2736a0e2d189f7d9700997a3c857a
refs/heads/master
2016-09-05T11:33:54.287376
2011-12-14T17:01:45
2011-12-14T17:01:45
40,006,358
1
0
null
null
null
null
UTF-8
Python
false
false
233
py
#!/usr/bin/env python3.1 import pickle data=('foo', 'bar', 'baz') with open('mydata', 'wb') as f: pickle.dump(data, f, pickle.HIGHEST_PROTOCOL) data=() with open('mydata', 'rb') as f: data = pickle.load(f) print(data)
[ "loris.tissino@5be9ef65-fe4e-d553-c31d-209ad9341544" ]
loris.tissino@5be9ef65-fe4e-d553-c31d-209ad9341544
dd58b30a002ad28ef364d7b282cc7d5afacbd9da
2a115b2a0a4d3c87ae6d262ba97a673a6c9a748b
/web-py/myWeb/get_img.py
741b804d6dfce4281c843bbc639d1882cda926de
[]
no_license
GoldenZhu/classroom
935407db720ecdda2c0ad05973a36784a205d0ab
26672b20477d3ae1d5aac9713302f1d3a6d1d834
refs/heads/master
2021-01-21T10:19:40.625219
2017-11-17T09:00:39
2017-11-17T09:00:45
83,407,700
0
1
null
null
null
null
UTF-8
Python
false
false
408
py
import urllib import time import web def get_poster(id, url): pic = urllib.urlopen(url).read() file_name = 'poster/%s.jpg' % id f = file(file_name, "wb") f.write(pic) f.close() db = web.database(dbn='sqlite', db='MovieSite.db') movies = db.select('movie') count = 0 for movie in movies: get_poster(movie.id, movie.image) count += 1 print count, movie.title time.sleep(2)
[ "2451255133@qq.com" ]
2451255133@qq.com
35365a1db34207810102ae5b1c2ee05f251e40f1
93c1d86cfa6893063519a89e73a647f0c2b48257
/stronka/migrations/0033_auto_20190617_1916.py
8e7099f79b5e85ec7f89d6a13685914ca616f883
[]
no_license
panchrobry/ProjektSzkieletoweREPO
24629649989a2eefe6af199e31b84e0b8d88b9b8
a0eeb244cfa8fc829739932ce167802b54f62382
refs/heads/master
2022-12-10T04:51:03.773554
2019-06-20T11:13:09
2019-06-20T11:13:09
177,784,319
0
0
null
2022-11-21T21:31:28
2019-03-26T12:29:41
Python
UTF-8
Python
false
false
453
py
# Generated by Django 2.0.13 on 2019-06-17 17:16 import datetime from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('stronka', '0032_auto_20190617_1737'), ] operations = [ migrations.AlterField( model_name='team', name='Register_Date', field=models.DateField(default=datetime.datetime(2019, 6, 17, 19, 16, 43, 700263)), ), ]
[ "karolek9.10@o2.pl" ]
karolek9.10@o2.pl
a07db48cd66c17dc6ba5bed0065585a071c5c8a4
8f83fb0541a9b542de4f4456bd2c14d6ed188b95
/mysite/config/settings/local.py
d8c13717c8331dda3aae85834587525058f6ad0e
[]
no_license
klee2017/Deploy-EB-docker
fc3b8c7d577f141b0273290f6fd8716f2e423204
b78292d6ef2f38987a1b6ad816ffaa89c4698789
refs/heads/master
2021-07-24T12:23:02.248951
2017-11-03T08:01:02
2017-11-03T08:01:02
108,818,956
0
0
null
null
null
null
UTF-8
Python
false
false
385
py
import random import string from .base import * ALLOWED_HOSTS = [ 'localhost', '127.0.0.1', '.elasticbeanstalk.com', '.locomoco.co.kr', ] DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': os.path.join(BASE_DIR, 'db.sqlite3'), } } SECRET_KEY = ''.join( [random.choice(string.ascii_lowercase) for i in range(40)] )
[ "kaythechemist@gmail.com" ]
kaythechemist@gmail.com
e4902d2ddf6c8dfca1b465dcd01596243d1635fb
063a45f56dab2a1ae77aaa3a96539fe25512e15c
/gazebo_turtlebot3_dqlearn.py
1af1209067bfd6c2a3096ab19a6e0700c931d28e
[]
no_license
changpowei/ROS_Gazebo_SB
44fe7c4312faad6b095a956e5074c26dfed25328
229759fd13f98d36e3802f4e4ac3ced19d4dab14
refs/heads/master
2023-08-23T21:34:22.277892
2021-10-21T05:38:49
2021-10-21T05:38:49
419,157,585
0
0
null
null
null
null
UTF-8
Python
false
false
18,481
py
import rospy import roslaunch import time import numpy as np import math import random from gazebo_msgs.srv import SpawnModel, DeleteModel, SetModelState from gazebo_msgs.msg import ModelState from geometry_msgs.msg import Twist from geometry_msgs.msg import Pose from geometry_msgs.msg import Point from nav_msgs.msg import Odometry import tf from sensor_msgs.msg import LaserScan from std_srvs.srv import Empty """ There are 3 different maze map in this packet After start one of them with launch file you have to edit this parameter for that maze. Options: maze1 maze2 maze3 """ SELECT_MAP = "maze1" class AgentPosController(): ''' This class control robot position We teleport our agent when environment reset So agent start from different position in every episode ''' def __init__(self): self.agent_model_name = "turtlebot3_waffle" def teleportRandom(self): ''' Teleport agent return new x and y point return agent posX, posY in list ''' model_state_msg = ModelState() model_state_msg.model_name = self.agent_model_name """Set the position of the begin""" if SELECT_MAP == "maze1": # maze 1 """ xy_list = [ [-1.5, 0.5], [-1.5, 1.5], [-0.5, 0.5], [-0.5, 1.5], [0.5, -0.5], [0.5, -1.5], [2.5, -0.5], [2.5, 0.5], [5.5,-1.5], [5.5,-0.5], [5.5,0.5], [5.5,1.5] ] """ xy_list = [ [-1.5, 0.5], [-1.5, 1.5], [-0.5, 0.5], [-0.5, 1.5], [0.5, -0.5], [0.5, -1.5], [2.5, -0.5], [2.5, 0.5], [5.5,-1.5], [5.5,-0.5], [5.5,0.5], [5.5,1.5] ] elif SELECT_MAP == "maze2": # maze 2 xy_list = [ [-1.5,-1.5], [-0.5,-1.5], [-1.5,-0.5], [-0.5,1.5], [1.5,0.5], [2.5,2.5], [2.5,3.5], [1.5,3.5], ] else: # maze 3 xy_list = [ [0.5,0.5], [1.5,0.5], [0.5,1.5], [1.5,1.5], [-0.5,-0.5], [-1.5,-0.5], [-1.5,-1.5], [0.5,-0.5], [0.5,-1.5], [1.5,-1.5], [-1.5,0.5], [-0.5,1.5], [-1.5,1.5], ] # Get random position for agent """ # A representation of pose in free space, composed of position and orientation. Point position Quaternion orientation """ pose = Pose() pose.position.x, pose.position.y = random.choice(xy_list) model_state_msg.pose = pose """ MSG: geometry_msgs/Point # This contains the position of a point in free space float64 x float64 y float64 z """ model_state_msg.twist = Twist() """ string reference_frame # set pose/twist relative to the frame of this entity (Body/Model) # leave empty or "world" or "map" defaults to world-frame """ model_state_msg.reference_frame = "world" # Start teleporting in Gazebo isTeleportSuccess = False for i in range(5): if not isTeleportSuccess: try: rospy.wait_for_service('/gazebo/set_model_state') telep_model_prox = rospy.ServiceProxy('/gazebo/set_model_state', SetModelState) telep_model_prox(model_state_msg) isTeleportSuccess = True break except Exception as e: rospy.logfatal("Error when teleporting agent " + str(e)) else: rospy.logwarn("Trying to teleporting agent..." + str(i)) time.sleep(2) if not isTeleportSuccess: rospy.logfatal("Error when teleporting agent") return "Err", "Err" return pose.position.x, pose.position.y class GoalController(): """ This class controls target model and position """ def __init__(self): self.model_path = "../models/gazebo/goal_sign/model.sdf" f = open(self.model_path, 'r') self.model = f.read() self.goal_position = Pose() self.goal_position.position.x = None # Initial positions self.goal_position.position.y = None self.last_goal_x = self.goal_position.position.x self.last_goal_y = self.goal_position.position.y self.model_name = 'goal_sign' self.check_model = False # This used to checking before spawn model if there is already a model def respawnModel(self): ''' Spawn model in Gazebo ''' isSpawnSuccess = False for i in range(5): if not self.check_model: # This used to checking before spawn model if there is already a model try: rospy.wait_for_service('gazebo/spawn_sdf_model') spawn_model_prox = rospy.ServiceProxy('gazebo/spawn_sdf_model', SpawnModel) spawn_model_prox(self.model_name, self.model, 'robotos_name_space', self.goal_position, "world") isSpawnSuccess = True self.check_model = True break except Exception as e: rospy.logfatal("Error when spawning the goal sign " + str(e)) else: rospy.logwarn("Trying to spawn goal sign ..." + str(i)) time.sleep(2) if not isSpawnSuccess: rospy.logfatal("Error when spawning the goal sign") def deleteModel(self): ''' Delete model from Gazebo ''' while True: if self.check_model: try: rospy.wait_for_service('gazebo/delete_model') del_model_prox = rospy.ServiceProxy('gazebo/delete_model', DeleteModel) del_model_prox(self.model_name) self.check_model = False break except Exception as e: rospy.logfatal("Error when deleting the goal sign " + str(e)) else: break def calcTargetPoint(self): """ This function return a target point randomly for robot """ self.deleteModel() # Wait for deleting time.sleep(0.5) if SELECT_MAP == "maze1": # maze 1 """ goal_xy_list = [ [-1.5, 0.5], [-1.5, 1.5], [-0.5, 0.5], [-0.5, 1.5], [0.5, -0.5], [0.5, -1.5], [2.5, -0.5], [2.5, 0.5], [5.5,-1.5], [5.5,-0.5], [5.5,0.5], [5.5,1.5] ] """ goal_xy_list = [ [-1.5, 0.5], [-1.5, 1.5], [-0.5, 0.5], [-0.5, 1.5], [0.5, -0.5], [0.5, -1.5], [2.5, -0.5], [2.5, 0.5], [5.5,-1.5], [5.5,-0.5], [5.5,0.5], [5.5,1.5] ] elif SELECT_MAP == "maze2": # maze 2 goal_xy_list = [ [-1.5,-1.5], [-0.5,-1.5], [-1.5,-0.5], [-0.5,1.5], [1.5,0.5], [2.5,2.5], [2.5,3.5], [1.5,3.5], ] else: # maze 3 goal_xy_list = [ [0.5,0.5], [1.5,0.5], [0.5,1.5], [1.5,1.5], [-0.5,-0.5], [-1.5,-0.5], [-1.5,-1.5], [0.5,-0.5], [0.5,-1.5], [1.5,-1.5], [-1.5,0.5], [-0.5,1.5], [-1.5,1.5], ] # Check last goal position not same with new goal while True: self.goal_position.position.x, self.goal_position.position.y = random.choice(goal_xy_list) if self.last_goal_x != self.goal_position.position.x: if self.last_goal_y != self.goal_position.position.y: break # Spawn goal model self.respawnModel() self.last_goal_x = self.goal_position.position.x self.last_goal_y = self.goal_position.position.y # Inform user rospy.logwarn("New goal position : " + str(self.goal_position.position.x) + " , " + str(self.goal_position.position.y)) return self.goal_position.position.x, self.goal_position.position.y def getTargetPoint(self): return self.goal_position.position.x, self.goal_position.position.y class Turtlebot3GymEnv(): ''' Main Gazebo environment class Contains reset and step function ''' def __init__(self): # Initialize the node rospy.init_node('turtlebot3_gym_env', anonymous=True) # Connect to gazebo self.velPub = rospy.Publisher('/cmd_vel', Twist, queue_size=5) self.unpause = rospy.ServiceProxy('/gazebo/unpause_physics', Empty) self.pause = rospy.ServiceProxy('/gazebo/pause_physics', Empty) self.reset_proxy = rospy.ServiceProxy( '/gazebo/reset_simulation', Empty) self.laserPointCount = 24 # 24 laser point in one time self.minCrashRange = 0.2 # Asume crash below this distance self.laserMinRange = 0.2 # Modify laser data and fix min range to self.laserMaxRange = 10.0 # Modify laser data and fix max range to self.stateSize = self.laserPointCount + 4 # Laser(arr), heading, distance, obstacleMinRange, obstacleAngle self.actionSize = 5 # Size of the robot's actions self.targetDistance = 0 # Distance to target self.targetPointX = 0 # Target Pos X self.targetPointY = 0 # Target Pos Y # Means robot reached target point. True at beginning to calc random point in reset func self.isTargetReached = True self.goalCont = GoalController() self.agentController = AgentPosController() def pauseGazebo(self): ''' Pause the simulation ''' rospy.wait_for_service('/gazebo/pause_physics') try: self.pause() except Exception: print("/gazebo/pause_physics service call failed") def unpauseGazebo(self): ''' Unpause the simulation ''' rospy.wait_for_service('/gazebo/unpause_physics') try: self.unpause() except Exception: print("/gazebo/unpause_physics service call failed") def resetGazebo(self): ''' Reset simualtion to initial phase ''' rospy.wait_for_service('/gazebo/reset_simulation') try: self.reset_proxy() except Exception: print("/gazebo/reset_simulation service call failed") def getLaserData(self): ''' ROS callback function return laser scan in 2D list ''' try: laserData = rospy.wait_for_message('/scan', LaserScan, timeout=5) return laserData except Exception as e: rospy.logfatal("Error to get laser data " + str(e)) def getOdomData(self): ''' ROS callback function Modify odom data quaternion to euler return yaw, posX, posY of robot known as Pos2D ''' try: odomData = rospy.wait_for_message('/odom', Odometry, timeout=5) odomData = odomData.pose.pose quat = odomData.orientation quatTuple = ( quat.x, quat.y, quat.z, quat.w, ) roll, pitch, yaw = tf.transformations.euler_from_quaternion( quatTuple) robotX = odomData.position.x robotY = odomData.position.y return yaw, robotX, robotY except Exception as e: rospy.logfatal("Error to get odom data " + str(e)) def calcHeadingAngle(self, targetPointX, targetPointY, yaw, robotX, robotY): ''' Calculate heading angle from robot to target return angle in float ''' targetAngle = math.atan2(targetPointY - robotY, targetPointX - robotX) heading = targetAngle - yaw if heading > math.pi: heading -= 2 * math.pi elif heading < -math.pi: heading += 2 * math.pi return round(heading, 2) def calcDistance(self, x1, y1, x2, y2): ''' Calculate euler distance of given two points return distance in float ''' return math.sqrt((x1 - x2)**2 + (y1 - y2)**2) def calculateState(self, laserData, odomData): ''' Modify laser data Calculate heading angle Calculate distance to target Calculate min range to nearest obstacle Calculate angle to nearest obstacle returns state as np.array State contains: laserData, heading, distance, obstacleMinRange, obstacleAngle ''' heading = self.calcHeadingAngle( self.targetPointX, self.targetPointY, *odomData) _, robotX, robotY = odomData distance = self.calcDistance( robotX, robotY, self.targetPointX, self.targetPointY) isCrash = False # If robot hit to an obstacle laserData = list(laserData.ranges) for i in range(len(laserData)): if (self.minCrashRange > laserData[i] > 0): isCrash = True if np.isinf(laserData[i]): laserData[i] = self.laserMaxRange if np.isnan(laserData[i]): laserData[i] = 0 obstacleMinRange = round(min(laserData), 2) obstacleAngle = np.argmin(laserData) return laserData + [heading, distance, obstacleMinRange, obstacleAngle], isCrash def step(self, action): ''' Act in envrionment After action return new state Calculate reward Calculate bot is crashed or not Calculate is episode done or not returns state as np.array State contains: laserData, heading, distance, obstacleMinRange, obstacleAngle, reward, done ''' self.unpauseGazebo() # Move maxAngularVel = 1.5 angVel = ((self.actionSize - 1)/2 - action) * maxAngularVel / 2 velCmd = Twist() velCmd.linear.x = 0.15 velCmd.angular.z = angVel self.velPub.publish(velCmd) # More basic actions """ if action == 0: #BRAKE LEFT velCmd = Twist() velCmd.linear.x = 0.17 velCmd.angular.z = 1.6 self.velPub.publish(velCmd) elif action == 1: #LEFT velCmd = Twist() velCmd.linear.x = 0.17 velCmd.angular.z = 0.8 self.velPub.publish(velCmd) elif action == 2: #FORWARD velCmd = Twist() velCmd.linear.x = 0.17 velCmd.angular.z = 0.0 self.velPub.publish(velCmd) elif action == 3: #RIGHT velCmd = Twist() velCmd.linear.x = 0.17 velCmd.angular.z = -0.8 self.velPub.publish(velCmd) elif action == 4: #BRAKE RIGHT velCmd = Twist() velCmd.linear.x = 0.17 velCmd.angular.z = -1.6 self.velPub.publish(velCmd) """ # Observe laserData = self.getLaserData() odomData = self.getOdomData() self.pauseGazebo() state, isCrash = self.calculateState(laserData, odomData) done = False if isCrash: done = True distanceToTarget = state[-3] if distanceToTarget < 0.2: # Reached to target self.isTargetReached = True if isCrash: reward = -150 elif self.isTargetReached: # Reached to target rospy.logwarn("Reached to target!") reward = 200 # Calc new target point self.targetPointX, self.targetPointY = self.goalCont.calcTargetPoint() self.isTargetReached = False else: # Neither reached to goal nor crashed calc reward for action yawReward = [] currentDistance = state[-3] heading = state[-4] # Calc reward # reference https://emanual.robotis.com/docs/en/platform/turtlebot3/ros2_machine_learning/ for i in range(self.actionSize): angle = -math.pi / 4 + heading + (math.pi / 8 * i) + math.pi / 2 tr = 1 - 4 * math.fabs(0.5 - math.modf(0.25 + 0.5 * angle % (2 * math.pi) / math.pi)[0]) yawReward.append(tr) try: distanceRate = 2 ** (currentDistance / self.targetDistance) except Exception: print("Overflow err CurrentDistance = ", currentDistance, " TargetDistance = ", self.targetDistance) distanceRate = 2 ** (currentDistance // self.targetDistance) reward = ((round(yawReward[action] * 5, 2)) * distanceRate) return np.asarray(state), reward, done def reset(self): ''' Reset the envrionment Reset bot position returns state as np.array State contains: laserData, heading, distance, obstacleMinRange, obstacleAngle ''' self.resetGazebo() while True: # Teleport bot to a random point agentX, agentY = self.agentController.teleportRandom() if self.calcDistance(self.targetPointX, self.targetPointY, agentX, agentY) > self.minCrashRange: break else: rospy.logerr("Reteleporting the bot!") time.sleep(2) if self.isTargetReached: while True: self.targetPointX, self.targetPointY = self.goalCont.calcTargetPoint() if self.calcDistance(self.targetPointX, self.targetPointY, agentX, agentY) > self.minCrashRange: self.isTargetReached = False break else: rospy.logerr("Recalculating the target point!") time.sleep(2) # Unpause simulation to make observation self.unpauseGazebo() laserData = self.getLaserData() odomData = self.getOdomData() self.pauseGazebo() state, isCrash = self.calculateState(laserData, odomData) self.targetDistance = state[-3] self.stateSize = len(state) return np.asarray(state) # Return state
[ "c95cpw@ncsist.org.tw" ]
c95cpw@ncsist.org.tw
bfc85210f3b259e11c7149a04537292c0612d22b
dfb193385a1d15f0aa0898ee67a7abd837666a5f
/RPi4B/dht22.py
eb036879e6ac282710090a6af7b1b70e8719d759
[]
no_license
OSHW-UHU-Group/SBA-IST
ed3a25e01c9efbc35262187c1b6f29f812138c9e
3c3a2cfafea0754f6aa51151db828401a7048ab6
refs/heads/master
2023-06-16T08:29:26.648085
2021-07-09T11:36:52
2021-07-09T11:36:52
379,353,309
1
0
null
null
null
null
UTF-8
Python
false
false
546
py
import time import board import adafruit_dht dht = adafruit_dht.DHT22(board.D26, use_pulseio=False) def humidity(): global dht i = 0 while i<10: try: temperature_c = dht.temperature humidity = dht.humidity return humidity, temperature_c except RuntimeError as error: # Reading doesn't always work! (try again max. 10) i = i + 1 time.sleep(1.0) continue # #while 1: # h, t= humidity() # print(h) # print(t) # time.sleep(1)
[ "alvarez@uhu.es" ]
alvarez@uhu.es
9670e8c23fe063f1c2f789cdbab33cb5eb371392
543efc7b32ab55a44ba4b85029c322228ac207db
/scripts/fit_aves.py
80a60cb3346d348755ab51ba454b7a6467e7899a
[]
no_license
hmnaik/aves
2238627b0a73237e74479f67eb78ee198283aab1
92cb2c5a50fd65d7056f93f209931e721946e818
refs/heads/main
2023-05-05T13:20:37.315266
2021-05-19T17:56:55
2021-05-19T17:56:55
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,381
py
import os import argparse import numpy as np import matplotlib.pyplot as plt import torch import _init_paths from models import mesh_regressor, AVES from optimization import base_renderer, AVES_Fitter from utils.renderer import Renderer from utils.cub_dataset import CUB_Dataset parser = argparse.ArgumentParser() parser.add_argument('--device', default='cuda', help='Device to use') parser.add_argument('--species_id', default=2, type=int, help='Species to run reconstruction') args = parser.parse_args() device = args.device species_id = args.species_id # dataset root = 'data/CUB' dataset = CUB_Dataset(root, species_id=species_id) images = dataset.images keypoints = dataset.keypoints segmentations = dataset.segmentations # objects size = 256 focal = 2167 aves = AVES(device=device, high_res=True) renderer = Renderer(focal, (size/2, size/2), img_w=size, img_h=size, faces=aves.dd['F']) silhouette_renderer = base_renderer(size=256, focal=2167, device=device) regressor = mesh_regressor(device=device) avesfit = AVES_Fitter(model=aves, prior_weight=10, mask_weight=1, beta_weight=150, global_iters=180, pose_iters=300, mask_iters=100, renderer=silhouette_renderer, device=device) # Regression to initialize print('Reconstructing', dataset.species, 'using AVES') print('Initializing ...') with torch.no_grad(): k = torch.tensor(keypoints).float().to(device) k[:, [9,15], :] = 0 k = k.reshape(-1, 54) pose, bone, tran = regressor(k) pose = regressor.postprocess(pose) # Optimize alignment print('Optimizing AVES ...') masks = torch.tensor(segmentations).clone().float().to(device) kpts = torch.tensor(keypoints).clone().float().to(device) pose_op, bone_op, tran_op, beta, model_mesh, model_kpts = avesfit(pose, bone, tran, focal_length=2167, camera_center=128, keypoints=kpts, masks=masks.squeeze(1), favor_mask=True) # Render and save all results print('Saving results ...') output_dir = 'output_aves_{}'.format(species_id) if not os.path.exists(output_dir): os.mkdir(output_dir) for i in range(len(images)): img = images[i] img_aves, _ = renderer(model_mesh[i].detach().cpu().numpy(), np.eye(3), [0,0,0], img) img_out = np.hstack([img, img_aves]) plt.imsave(output_dir + '/{:04d}.png'.format(i), img_out.astype(np.uint8))
[ "yufu@seas.upenn.edu" ]
yufu@seas.upenn.edu
9660216677c127738ccae93efffd11f88bb9cf9a
fa21ed4e665f00bd22236b6b224b247014ff2772
/nodes/speed_controller.py
a964d53f56b5d404997bc813aff01b95e3d68cdf
[]
no_license
klipfel/rosGoturn
87d77969102e22aff94ce83a07c77556d0e47062
2243259820a542791df5baed0eb97c16836dba46
refs/heads/master
2020-06-15T23:00:27.008175
2019-07-19T13:45:44
2019-07-19T13:45:44
195,414,798
0
0
null
null
null
null
UTF-8
Python
false
false
1,803
py
#!/usr/bin/env python3 import board import adafruit_pca9685 import busio import time import os class Servomotor: def __init__(self): i2c = busio.I2C(board.SCL, board.SDA) self.pwm = adafruit_pca9685.PCA9685(i2c) self.pwm.frequency = 50 # values self.leftMax = 100 self.rightMax = 0 self.straight = 50 self.angle = None self.set_bearing(self.straight) print("Servomotor initialization SUCCESS") def test(self): accuracy = 10 for angle in range(0,100 + accuracy,accuracy): self.set_bearing(angle) time.sleep(1) self.set_bearing(self.straight) def terminal_test(self): value = input("Angle [0-100]: ") while value != 'q': self.set_bearing(float(value)) value = input("Angle [0-100]: ") print("value entered : " + value) def set_bearing(self,angle): self.pwm.channels[1].duty_cycle = int(3932+ angle*2620/100) self.angle = angle class Motor: def __init__(self): i2c = busio.I2C(board.SCL, board.SDA) self.pwm = adafruit_pca9685.PCA9685(i2c) self.pwm.frequency = 50 # values self.off = 50 self.forwardMin = 68 # motor setup self.setup() print("Motor initialization SUCCESS") def stop(self): self.set_speed(self.off) def setup(self): self.set_speed(50) time.sleep(1) def test(self): for speed in range(0,100,10): self.set_speed(speed) time.sleep(1) self.set_speed(self.forwardMin) time.sleep(4) self.stop() def terminal_test(self): value = input("Speed [0-100]: ") while value != 'q': self.set_speed(float(value)) value = input("Speed [0-100]: ") print("value entered : " + value) def set_speed(self,speed): self.pwm.channels[0].duty_cycle = int(3932+ speed*2620/100) if __name__ == "__main__": print("test") b = Servomotor() m = Motor() b.test() m.terminal_test()
[ "arnaudklipfel@hotmail.com" ]
arnaudklipfel@hotmail.com
1b006e7f6b992e3c1dccc846fe3689745833841a
9c88bd2c194ccc19b4989bf844564c5857e2ea82
/Simple_Calculations/rectangle_area.py
379677136a28e027dd00a9629978df03cbd0f470
[]
no_license
ectky/PythonProjects
faa1d600f6b1e3aaee19b96a1cbf384ac0dc47a2
5920ee1752df729c178fe088b58ba17d8585b333
refs/heads/master
2021-08-11T09:31:38.562054
2017-11-13T13:59:04
2017-11-13T13:59:04
110,552,837
0
0
null
null
null
null
UTF-8
Python
false
false
281
py
import math x1 = float(input()) y1 = float(input()) x2 = float(input()) y2 = float(input()) upper_side = math.fabs(x1 - x2) right_side = math.fabs(y2 - y1) area = upper_side * right_side perimeter = upper_side * 2 + right_side * 2 print(area) print(perimeter)
[ "noreply@github.com" ]
noreply@github.com
49db5b641d88aa13b1b34b6fcceea02798bae5a6
cf64c598ee8ace5eb72a3992ca214b8f05383177
/Tools/lib/cherrysoda.py
09e40be6062bab2616daef95af54740f9d9178ca
[ "MIT" ]
permissive
brucelevis/cherrysoda-engine
6c6efb9545a27ca8a612b56e6ffb61a8347a3923
64b2ff5a2a53831ded3e60ff2a28a101fa84f3fd
refs/heads/master
2023-02-01T15:57:38.809819
2020-12-17T02:35:01
2020-12-17T02:35:01
null
0
0
null
null
null
null
UTF-8
Python
false
false
3,933
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import glob import os import pathlib import platform import shutil import subprocess import sys import zipfile if sys.version_info.major == 3: import urllib.request as urllib2 else: import urllib2 def is_windows_system(): return platform.system() == 'Windows' def join_path(a, b): return os.path.join(a, b) def abspath(a): return os.path.abspath(a) def get_file_path(f): return os.path.dirname(f) def get_file_name(f): return os.path.basename(f) executable_suffix = ('', '.exe')[is_windows_system()] project_path = abspath(join_path(get_file_path(__file__), '../..')) engine_path = join_path(project_path, 'Engine') tool_path = join_path(project_path, 'Tools') external_path = join_path(project_path, 'External') tmp_path = join_path(project_path, 'Tmp') bgfx_src_path = join_path(external_path, 'bgfx/bgfx/src') shaderc = join_path(tool_path, 'bin/shaderc' + executable_suffix) sdl2_version = '2.0.12' sdl2_path = join_path(external_path, 'SDL2-' + sdl2_version) def make_sure_folder_exist(f): p = os.path.dirname(f) pathlib.Path(p).mkdir(parents=True, exist_ok=True) def execute_command(command): # print('$ ' + ' '.join(command)) subprocess.run(command) def compile_shader(shader_source, output, platform, shader_type, include_path=None, profile=None, opt_level=None, bin2c_array=None): command = [shaderc, '-f', shader_source, '-o', output, '--platform', platform, '--type', shader_type] if include_path: command += ['-i', include_path] if profile: command += ['--profile', profile] if opt_level: command += ['-O', str(opt_level)] if bin2c_array: command += ['--bin2c', bin2c_array] make_sure_folder_exist(output) execute_command(command) def download_url_to(url, dest): filename = url.split('/')[-1] print("Downloading '%s' from '%s'..." % (filename, url)) u = urllib2.urlopen(url) filedest = os.path.join(dest, filename) make_sure_folder_exist(filedest) f = open(filedest, 'wb') f.write(u.read()) f.close() print("Finished!") return filedest def extract_zip_to(filename, dest): zfile = zipfile.ZipFile(filename, 'r') for ef in zfile.namelist(): if ef.endswith('/'): continue p = os.path.join(dest, ef) make_sure_folder_exist(p) f = open(p, 'wb') f.write(zfile.read(ef)) f.close() zfile.close() def get_file_list_from_wildcard(wildcard): return glob.glob(wildcard) def get_file_list_of_path(path): walkList = os.walk(path) file_list = [] for i in walkList: root, dirs, files = i for f in files: file_list.append(os.path.join(root, f)) return file_list def exists(path): return os.path.exists(path) def copy(src, dest): src_list = get_file_list_from_wildcard(src) for source in src_list: shutil.copy(source, dest) def copytree(src, dest): shutil.copytree(src, dest) def move(src, dest): src_list = get_file_list_from_wildcard(src) for source in src_list: shutil.move(source, dest) def set_environment_variable(env_var, value): command = ['setx', env_var, value] execute_command(command) def write_str_file(s, dest): f = open(dest, 'w') f.write(s) f.close() def read_file(file): f = open(file, 'r') ret = f.read() f.close() return ret def replace_file_str(file, replace_list): s = read_file(file) for i in replace_list: find, rep = i s = s.replace(find, rep) write_str_file(s, file) def replace_file_name(file, replace_list): s = file for i in replace_list: find, rep = i s = s.replace(find, rep) if file != s: if exists(s): print('"%s" already exists!' % (s)) return move(file, s)
[ "c.even@live.cn" ]
c.even@live.cn
0aac049c8263f7e956cea14027ed8e142b6344e5
0931696940fc79c4562c63db72c6cabfcb20884d
/Exercises/Regular_Expresions/furniture.py
8a02f7b386384bfbe0d6b9fe2cf832c3b0cd53d3
[]
no_license
ivklisurova/SoftUni_Fundamentals_module
f847b9de9955c8c5bcc057bb38d57162addd6ad8
69242f94977c72005f04da78243a5113e79d6c33
refs/heads/master
2021-12-01T01:56:22.067928
2021-11-08T17:07:31
2021-11-08T17:07:31
253,281,893
0
0
null
null
null
null
UTF-8
Python
false
false
468
py
import re furniture = [] total_money = 0 while True: order = input() if order == 'Purchase': break pattern = r'>{2}([a-zA-z]+)<{2}(\d+[.]\d+|\d+)!(\d+)' matches = re.findall(pattern, order) for i in matches: if len(i) == 0: break furniture.append(i[0]) total_money += float(i[1]) * float(i[2]) print('Bought furniture:') [print(x) for x in furniture] print(f'Total money spend: {total_money:.2f}')
[ "55747390+ivklisurova@users.noreply.github.com" ]
55747390+ivklisurova@users.noreply.github.com
f3c4d9fd7986ae013086ca33b92b7e7b2de6fac1
831b2c2d88d03ce366c091ebc851281a20b3c567
/clean_unsupervised/predictorUnsupervised.py
18398a5162c9cd84ea48ea8cc360a732834bfba4
[]
no_license
nathanieljblack/PriceRight
cb2d0ca0929f6b5791397bd1f72267da20ac0144
c4fb112ae27f3adbd75bb88603d23db35de29ff3
refs/heads/master
2020-04-19T11:24:56.683210
2015-05-01T14:05:33
2015-05-01T14:05:33
67,873,785
1
0
null
null
null
null
UTF-8
Python
false
false
13,011
py
import pandas as pd import re from unicodedata import category as cat from nltk.corpus import stopwords from gensim import corpora, models import operator from collections import OrderedDict import scipy from scipy.cluster.vq import kmeans, vq class Predictor(): # Note: This predictor trains the model and runs the model on the same data set def __init__(self, list): """ :param list: a list of dictionaries as read in from a json file """ self.data = pd.DataFrame(list) def regularize_training_corpus(self): def clean_titles(title): if re.search('(wtb|wanted|want|purchase|repair|buy|need|trade|replacement| \ looking|fix|cash|me|pawn|wtt|trading)', title, re.IGNORECASE): out = "exclude" # common non-sales words elif re.search('(ipad|ipod)', title, re.IGNORECASE): out = "exclude" # common iPhone related items for sale else: out = "OK" return out print "\nRegularizing titles ..." self.data['status'] = self.data.title.apply(clean_titles) self.data = self.data[self.data.status != 'exclude'] return self.data def tokenize_training_corpus(self): stoplist = stopwords.words('english') sales_words = ['like', 'new', 'brand', 'excellent', 'condition', 'pristine', 'never', 'used', 'clean', 'perfect', 'great', 'sale', 'sell', 'selling', 'good', 'obo', 'warranty', 'cl', 'color', 'works', 'extras', 'open', 'flawless', 'bad', 'guaranteed', 'working', 'unopened' ] # common sales words iphone_words = ['iphone', 'apple', 'icloud', 'contract', 'iphones', 'carrier', 'verizon', 'tmobile', 'att', 'metropcs', 'sprint', 'cricket', 'wireless', 'mobile', 'phone', 'smartphone', 'unlocked', 'unlock', 'locked', 'lock', 'factory', 'box', 'sealed', 'gsm', 'esn', '4g', 'imei', 'international', 'cracked', 'screen', 'charger', '8', '16', '32', '64', '128', 'g', '8g', '16g', '32g', '64g', '128g', 'gb', '8gb', '16gb', '32gb', '64gb', '128gb', 'gig', '8gig', '16gig', '32gig', '64gig','128gig', 'gigs', '8gigs', '16gigs', '32gigs', '64gigs','128gigs', 'white', 'black', 'gray', 'grey', 'spacegray', 'spacegrey', 'space', 'pink', 'mint', 'gold', 'silver', 'blue', 'yellow', 'green', 'pink', 'whitesilver', 'blackgray', 'slate', 'whitegold' ] # words that will appear across all iphone brands (4, 4s, 5, 5c, 5s, 6, 6+) custom_stoplist = sales_words + iphone_words # Tokenize titles def create_tokens(title): out = [] for word in title.lower().split(): out.append(word) return out # Remove punctuation def strip_punctuation(token): out = [] for word in token: if __name__ == "__main__": new_word = "".join(char for char in word if not cat(char).startswith('P')) else: new_word = "".join(char for char in word.decode('utf-8') if not cat(char).startswith('P')) out.append(new_word) return out # Remove common words def remove_common_words(token): out = [] for word in token: if word not in stoplist and word not in custom_stoplist and word != '': out.append(word) return out # Remove words that appear only once def remove_once_words(token): out = [] for word in token: if word not in tokens_once: out.append(word) return out # Exclude postings whose tokens are empty def remove_empty_tokens(token): if token: out = "OK" else: out = "exclude" return out def cheat_replace(title): newtitle = title.replace('6 plus', '6+') # Cheat newtitle = newtitle.replace('6 Plus', '6+') # Cheat newtitle = newtitle.replace('6 PLUS', '6+') # Cheat newtitle = newtitle.replace('6plus', '6+') # Cheat newtitle = newtitle.replace('6Plus', '6+') # Cheat newtitle = newtitle.replace('6PLUS', '6+') # Cheat return newtitle print "Preparing Training Corpus ..." print " Cheating ..." self.data.title = self.data.title.apply(cheat_replace) print " Creating tokens ..." self.data['tokens'] = self.data.title.apply(create_tokens) print " Removing punctuation ..." self.data.tokens = self.data.tokens.apply(strip_punctuation) print " Removing common words ..." self.data.tokens = self.data.tokens.apply(remove_common_words) print " Removing words that appear only once ..." all_tokens = sum(self.data.tokens, []) tokens_once = set(word for word in set(all_tokens) if all_tokens.count(word) == 1) self.data.tokens = self.data.tokens.apply(remove_once_words) print " Removing empty tokens ..." self.data.status = self.data.tokens.apply(remove_empty_tokens) self.data = self.data[self.data.status != 'exclude'] return self.data def train_model(self, numTopics): # Create training dictionary def create_dictionary(training_data): dictionary = corpora.Dictionary(training_data) dictionary.save('training_dictionary.dict') # store to disk, for later use return 'training_dictionary.dict' # Turn training corpus into sparse Bag of Words vectors def corpus_to_bag_of_words(training_dict, training_data): training_corpus_bow = [training_dict.doc2bow(token) for token in training_data] corpora.MmCorpus.serialize('training_corpus.mm', training_corpus_bow) # store to disk, for later use return 'training_corpus.mm' # Train TF-IDF model def corpus_to_tfidf(training_corp): tfidf_model = models.TfidfModel(training_corp) corpus_tfidf = tfidf_model[training_corp] return corpus_tfidf, tfidf_model # Train LSI model def corpus_to_lsi(dict, corpus, topics): lsi_model = models.LsiModel(corpus, id2word=dict, num_topics=topics) # initialize an LSI transformation topicWordProbMat = lsi_model.print_topics(topics) return lsi_model, topicWordProbMat def create_topic_dict(topic_words): def create_tokens(topic): out = [] for word in topic.lower().split(): out.append(word) return out topics_list = [] for topic in topic_words: topic_wds = [] topic = topic.replace(' + ', ' ') topic = topic.replace('*', ' ') topic = topic.replace('"', '') topic_tokens = create_tokens(topic) for i in range(1, len(topic_tokens)/2): pair = (float(topic_tokens[2*i-2]), topic_tokens[2*i-1],) topic_wds.append(pair) topics_list.append(topic_wds) for i in range(len(topics_list)): topics_list[i] = sorted(topics_list[i], key=operator.itemgetter(0)) topics_dict = OrderedDict() i = 0 for topic in topics_list: topics_dict[i] = dict(enumerate(topic)) i += 1 return topics_dict print "Training model ..." print " Creating training dictionary ..." training_dictionary_file = create_dictionary(self.data.tokens) training_dictionary = corpora.Dictionary.load(training_dictionary_file) print " Transforming training corpus into bag-of-words vectors..." training_corpus_file = corpus_to_bag_of_words(training_dictionary, self.data.tokens) training_corpus = corpora.MmCorpus(training_corpus_file) print " Creating TF-IDF vectors ..." corpus_tfidf, tfidf_model = corpus_to_tfidf (training_corpus) print " Training LSI model using " + str(numTopics) + " topics ..." lsi_model, topic_words = corpus_to_lsi(training_dictionary, corpus_tfidf, numTopics) print " Creating topic dictionary ..." topics_dict = create_topic_dict(topic_words) return self.data, training_dictionary, tfidf_model, lsi_model, topics_dict def run_model(self, training_dictionary, tfidf_model, lsi_model, num_topics, num_clusters): # Create LSI vectors for Clustering def create_lsi_vectors(token, dict, tfidf_model, lsi_model): vec_bow = dict.doc2bow(token) vec_tfidf = tfidf_model[vec_bow] # convert the token to TF-IDF space vec_lsi = lsi_model[vec_tfidf] # convert the token to LSI space return vec_lsi # Clean LSI vectors def clean_lsi_vectors(lsi_vectors, tops): # Remove vectors if they have less than numTopics elements def remove_short_vectors(vec): if len(vec) < tops: out = "exclude" else: out = "OK" return out # Check that LSI model created vectors of proper length print " Checking for short LSI vectors ..." minLength = tops + 100 maxLength = 0 numTooSmall = 0 for vector in lsi_vectors: if len(vector) < minLength: minLength = len(vector) if len(vector) > maxLength: maxLength = len(vector) if len(vector) < tops: numTooSmall +=1 print (" MinLength = " + str(minLength) +"\n MaxLength = " + str(maxLength)) if numTooSmall > 0: # if lsi model fails, remove short vectors print " After running the LSI model, " + str(numTooSmall) + " vectors were too short." print " Removing short LSI vectors ..." self.data.status = self.data.lsiVectors.apply(remove_short_vectors) self.data = self.data[self.data.status != 'exclude'] return self.data def cluster(numClusters): # Prep LSI vectors for clustering self.data.clusterVectors = [[x[1] for x in vector] for vector in self.data.lsiVectors] self.data.lsiArray = scipy.array(self.data.clusterVectors) # Compute K-Means print " Running K-Means clustering with " + str(numClusters) + " clusters ..." centroids, _ = kmeans(self.data.lsiArray, numClusters) # Assign each title to a cluster print " Assigning postings to their clusters ..." self.data['pred_bin'], _ = vq(self.data.lsiArray,centroids) # Save centroids print " Saving centroids ..." centroids_list = centroids.tolist() centroids_dict = OrderedDict() i = 0 for centroid in centroids_list: centroids_dict[i] = dict(enumerate(centroid)) i += 1 for i in range(len(centroids_list)): centroids_dict[i] = sorted(centroids_dict[i].items(), key=operator.itemgetter(1)) return self.data, centroids_dict print"Running model ..." print " Creating LSI vectors for clustering ..." self.data['lsiVectors'] = self.data.tokens.apply(create_lsi_vectors, args=(training_dictionary, tfidf_model, lsi_model)) print " Cleaning LSI vectors ..." self.data = clean_lsi_vectors(self.data.lsiVectors, num_topics) print" Clustering postings ..." self.data, centroids = cluster(num_clusters) return self.data, centroids # Create predictions based on the model def predict(self, numTopics, numClusters): self.data = self.regularize_training_corpus() self.data = self.tokenize_training_corpus() self.data, training_dictionary, tfidf_model, lsi_model, topics_dict = self.train_model(numTopics) self.data, centroids = self.run_model(training_dictionary, tfidf_model, lsi_model, numTopics, numClusters) self.data = self.data.drop(['status', 'lsiVectors'], axis=1) # keep tokens for now return self.data, topics_dict, centroids
[ "marguerite@oneto.us" ]
marguerite@oneto.us
7a5ddbcad3a15a9946b3546a3153387a9b875871
202e657b5c9bfcf3040146f779eafaf453d43d10
/number-of-people.py
cd46b5b272560c65fca28b84aae50f97b45be1bd
[]
no_license
derekmcauley7/star-wars-api-college-assignment
4695ff5c7393cefeb0d837a11c065431f3266998
1d080dbe7dc759de1f313fe727518e7e4feb8e6c
refs/heads/master
2020-05-09T17:05:50.835601
2019-04-14T11:08:44
2019-04-14T11:08:44
null
0
0
null
null
null
null
UTF-8
Python
false
false
733
py
from urllib.request import Request, urlopen from json import loads import ssl # How many people are returned by the API? Show how you can solve this without using the results count attribute?s certsPath='/usr/local/lib/python3.7/site-packages/pip/_vendor/certifi/cacert.pem' context=ssl.SSLContext() context.load_verify_locations(certsPath) url = 'http://swapi.co/api/people' results = [] while url != None: req = Request(url, None, { 'User-agent' : 'Mozilla/5.0 (Windows; U; Windows NT 5.1; de; rv:1.9.1.5) Gecko/20091102 Firefox/3.5.5' }) data = loads(urlopen(req, context=context).read().decode("utf-8")) results += data['results'] url = data['next'] print("Number of people: " + str(len(results)))
[ "derekmcauley7@gmail.com" ]
derekmcauley7@gmail.com
03d97741b6759ae1f367e5a530b53d4f599f17f5
ec841eb43a9bdf840e227bed932900405a89fe95
/maze.py
0f037d255b655f43d76b449720d3ac77c2d3f91d
[]
no_license
130e/SnakeGame
e358c85a0f80eb1740d75e93857eba8828b91aa1
28166ea346b7aae9728fa3f12355a1bf35733ab6
refs/heads/master
2022-11-17T13:02:14.806785
2020-07-16T18:19:19
2020-07-16T18:19:19
139,757,630
0
0
null
null
null
null
UTF-8
Python
false
false
4,902
py
# fernival made # set certain macros KEY_EXIT = 27 # define macros KEY_SPACE = 32 KEY_UP = 119 # WASD KEY_DOWN = 115 KEY_LEFT = 97 KEY_RIGHT = 100 KEY_LOST = KEY_DOWN NO_ENTRY = float('-inf') ABSORB = 1.0 MIN_TOLERANCE = 0.0000000001 # 10^-11 class node(): def __init__(self, y, x, value=0.0): self.pi = {KEY_UP: 0.25, KEY_DOWN: 0.25, KEY_LEFT: 0.25, KEY_RIGHT: 0.25} self.value = value self.neighborV = {KEY_UP: 0.0, KEY_DOWN: 0.0, KEY_LEFT: 0.0, KEY_RIGHT: 0.0} self.y = y self.x = x class maze(): def __init__(self, height, width, value=0.0): self.axis = (height, width) self.map = [[node(y, x, value=value) for y in range(width)] for x in range(height)] # (height, width) so map[height][width] def setAbsorb(self, cord): self.map[cord[0]-1][cord[1]-1].value = ABSORB def __refreshNeighborValue(self): for j in range(self.axis[1]): for i in range(self.axis[0]): if j == 0: self.map[i][j].neighborV[KEY_LEFT] = self.map[i][self.axis[1] - 1].value else: self.map[i][j].neighborV[KEY_LEFT] = self.map[i][j - 1].value if j == self.axis[1] - 1: self.map[i][j].neighborV[KEY_RIGHT] = self.map[i][0].value else: self.map[i][j].neighborV[KEY_RIGHT] = self.map[i][j + 1].value if i == 0: self.map[i][j].neighborV[KEY_UP] = self.map[self.axis[0] - 1][j].value else: self.map[i][j].neighborV[KEY_UP] = self.map[i - 1][j].value if i == self.axis[0] - 1: self.map[i][j].neighborV[KEY_DOWN] = self.map[0][j].value else: self.map[i][j].neighborV[KEY_DOWN] = self.map[i + 1][j].value def refreshValue(self, gamma=1, instantValue=-1): self.__refreshNeighborValue() bConverge = True # for j in range(self.axis[1]): # for i in range(self.axis[0]): # nd = self.map[i][j] # nd.value = nd.pi[KEY_UP] * (nd.value + nd.neighborV[KEY_UP]) + \ # nd.pi[KEY_DOWN] * (nd.value + nd.neighborV[KEY_DOWN]) + \ # nd.pi[KEY_LEFT] * (nd.value + nd.neighborV[KEY_LEFT]) + \ # nd.pi[KEY_RIGHT] * (nd.value + nd.neighborV[KEY_RIGHT]) for col in self.map: for nd in col: if nd.value == ABSORB or nd.value == NO_ENTRY: continue oldValue = nd.value nd.value = 0 for k in nd.pi.keys(): if nd.pi[k] == 0: # exclude cal with -inf continue nd.value += nd.pi[k] * (instantValue + gamma * nd.neighborV[k]) if bConverge == True: # check if converged if abs(oldValue - nd.value) >= MIN_TOLERANCE: bConverge = False return bConverge def refreshPi(self): self.__refreshNeighborValue() for col in self.map: for nd in col: moves = 0 for k in nd.pi.keys(): if nd.pi[k] != 0 and nd.neighborV[k] != NO_ENTRY and nd.neighborV[k] >= nd.value: moves += 1 else: nd.pi[k] = 0 if moves != 0: p = 1 / moves else: p = 0 for k in nd.pi.keys(): if nd.pi[k] != 0: nd.pi[k] = p # # no choice set the biggest, if surrounded, go up and die # max = float('-inf') # key = KEY_UP # for k in nd.pi.keys(): # if nd.neighborV[k] > max: # max = nd.neighborV[k] # key = k # for k in nd.pi.keys(): # nd.pi[k] = 0 # nd.pi[key] = 1 def refreshDead(self, cords): for c in cords: self.map[c[0]-1][c[1]-1].value = NO_ENTRY def getMove(self, cord): nd = self.map[cord[0]-1][cord[1]-1] max = nd.neighborV[KEY_DOWN] key = KEY_DOWN for k in nd.neighborV.keys(): if nd.pi[k] != 0 and max < nd.neighborV[k]: max = nd.neighborV[k] key = k return key def show(self): for i in range(self.axis[0]): l = [] for j in range(self.axis[1]): l.append(self.map[i][j].value) print(l)
[ "noreply@github.com" ]
noreply@github.com
e7f1ba699ec76f92212d50d4d70c0fb5f85ae38d
97228f81e03cebd8c250c267b4943bddb3640fd1
/gdksite/event/migrations/0011_auto_20181221_1244.py
e01af6b8008dd2e6ca92ca49f5c67484b84daf3c
[ "MIT" ]
permissive
vgrivtsov/moygdk
88e48e234d8737e044dc6bd9c4e1794c2a25fe32
74d63299c7326bcae92a17b61d978ad91f0d5552
refs/heads/master
2022-12-09T18:38:28.731819
2021-02-12T10:20:45
2021-02-12T10:20:45
164,198,744
0
0
MIT
2022-12-08T01:34:21
2019-01-05T09:02:21
JavaScript
UTF-8
Python
false
false
443
py
# Generated by Django 2.1.4 on 2018-12-21 12:44 from django.db import migrations class Migration(migrations.Migration): dependencies = [ ('event', '0010_auto_20181220_0821'), ] operations = [ migrations.RemoveField( model_name='eventpage', name='event_date', ), migrations.RemoveField( model_name='eventpage', name='event_time', ), ]
[ "vgrivtsov@gmail.com" ]
vgrivtsov@gmail.com
61f623bb2311199c6f90a06eafc6177b8604e7b1
a38856315e9a35f5eb0905a10eae6840741c468a
/stix_edh/cyber_profile.py
a1b921a93c6da80b797c6892d9627ef92aadfe44
[ "BSD-3-Clause" ]
permissive
emmanvg/stix-edh
bbf4cebb908ad8a7c7dd8728ebfc67284f17365d
b426f9785339ab741bb9fb21d356b36193791afc
refs/heads/master
2020-04-11T23:35:44.934139
2018-08-01T16:16:15
2018-08-01T16:16:15
162,172,740
0
0
NOASSERTION
2018-12-17T18:22:40
2018-12-17T18:22:39
null
UTF-8
Python
false
false
4,224
py
# Copyright (c) 2017, The MITRE Corporation. All rights reserved. # See LICENSE.txt for complete terms. # python-stix import stix from mixbox import fields # internal bindings from stix_edh.bindings import cyber_profile class AccessPrivilege(stix.Entity): _binding = cyber_profile _binding_class = _binding.AccessPrivilegeType _namespace = 'urn:edm:edh:cyber:v3' privilege_action = fields.TypedField("privilegeAction", type_="stix_edh.common.NMTokens", key_name="privilege_action") privilege_scope = fields.TypedField("privilegeScope", type_="stix_edh.common.NMTokens", multiple=True, key_name="privilege_scope") rule_effect = fields.TypedField("ruleEffect", type_="stix_edh.common.NMTokens", key_name="rule_effect") def __init__(self): super(AccessPrivilege, self).__init__() def add_privilege_scope(self, value): from stix_edh import common if not value: return nmtokens = common.NMTokens(value) self.privilege_scope.append(nmtokens) class ResourceDisposition(stix.Entity): _binding = cyber_profile _binding_class = _binding.ResourceDispositionType _namespace = 'urn:edm:edh:cyber:v3' disposition_date = fields.DateField("dispositionDate", key_name="disposition_date") disposition_process = fields.TypedField("dispositionProcess", type_="stix_edh.common.NMTokens", key_name="disposition_process") def __init__(self): super(ResourceDisposition, self).__init__() class OriginalClassification(stix.Entity): _binding = cyber_profile _binding_class = _binding.OriginalClassificationType _namespace = 'urn:edm:edh:cyber:v3' classified_by = fields.TypedField("classifiedBy", type_="stix_edh.common.NMTokens", key_name="classified_by") classified_on = fields.DateField("classifiedOn", key_name="classified_on") classification_reason = fields.TypedField("classificationReason", type_="stix_edh.common.NMTokens", key_name="classification_reason") compilation_reason = fields.TypedField("compilationReason", type_="stix_edh.common.NMTokens", key_name="compilation_reason") def __init__(self): super(OriginalClassification, self).__init__() class DerivativeClassification(stix.Entity): _binding = cyber_profile _binding_class = _binding.DerivativeClassificationType _namespace = 'urn:edm:edh:cyber:v3' classified_by = fields.TypedField("classifiedBy", type_="stix_edh.common.NMTokens", key_name="classified_by") classified_on = fields.DateField("classifiedOn", key_name="classified_on") derived_from = fields.TypedField("derivedFrom", type_="stix_edh.common.NMTokens", key_name="derived_from") def __init__(self): super(DerivativeClassification, self).__init__() class FurtherSharing(stix.Entity): _binding = cyber_profile _binding_class = _binding.FurtherSharingType _namespace = "urn:edm:edh:cyber:v3" rule_effect = fields.TypedField("ruleEffect", key_name="rule_effect") sharing_scope = fields.TypedField("sharingScope", type_="stix_edh.common.NMTokens", key_name="sharing_scope") def __init__(self): super(FurtherSharing, self).__init__() class Declassification(stix.Entity): _binding = cyber_profile _binding_class = _binding.DeclassificationType _namespace = 'urn:edm:edh:cyber:v3' declass_exemption = fields.TypedField("declassExemption", type_="stix_edh.common.NMTokens", key_name="declass_exemption") declass_period = fields.IntegerField("declassPeriod", key_name="declass_period") declass_date = fields.DateField("declassDate", key_name="declass_date") declass_event = fields.TypedField("declassEvent", type_="stix_edh.common.NMTokens", key_name="declass_event") def __init__(self): super(Declassification, self).__init__() class PublicRelease(stix.Entity): _binding = cyber_profile _binding_class = _binding.PublicReleaseType _namespace = 'urn:edm:edh:cyber:v3' released_by = fields.TypedField("releasedBy", type_="stix_edh.common.NMTokens", key_name="released_by") released_on = fields.DateField("releasedOn", key_name="released_on") def __init__(self): super(PublicRelease, self).__init__()
[ "gback@mitre.org" ]
gback@mitre.org
ff7fa112a3352d67dbc4074aac0cf24ea3e98617
7d35fcbcceb1ff4d458cef69a6eda3c5a3a5734b
/app/core/migrations/0001_initial.py
6a57eddf491c01619605b9ae2fe151f53636b5b4
[ "MIT" ]
permissive
hemant-mehra/UDEMY_REST_API_advance_course
4b1ea47037633090b06c037a8abc63eea73af8a7
c8c0298e3650e64d2fb35370fb5b8a1bae741937
refs/heads/main
2023-04-23T09:03:17.131445
2021-04-29T12:28:23
2021-04-29T12:28:23
361,781,723
0
0
null
null
null
null
UTF-8
Python
false
false
1,710
py
# Generated by Django 3.2 on 2021-04-27 15:24 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ('auth', '0012_alter_user_first_name_max_length'), ] operations = [ migrations.CreateModel( name='User', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('password', models.CharField(max_length=128, verbose_name='password')), ('last_login', models.DateTimeField(blank=True, null=True, verbose_name='last login')), ('is_superuser', models.BooleanField(default=False, help_text='Designates that this user has all permissions without explicitly assigning them.', verbose_name='superuser status')), ('email', models.EmailField(max_length=255, unique=True)), ('name', models.CharField(max_length=255)), ('is_active', models.BooleanField(default=True)), ('is_staff', models.BooleanField(default=False)), ('groups', models.ManyToManyField(blank=True, help_text='The groups this user belongs to. A user will get all permissions granted to each of their groups.', related_name='user_set', related_query_name='user', to='auth.Group', verbose_name='groups')), ('user_permissions', models.ManyToManyField(blank=True, help_text='Specific permissions for this user.', related_name='user_set', related_query_name='user', to='auth.Permission', verbose_name='user permissions')), ], options={ 'abstract': False, }, ), ]
[ "hemantmehra.p@gmail.com" ]
hemantmehra.p@gmail.com
bdc06186b30b43623dc40611426d6798cf09e14e
08c3f4b38204609d9f398f4b0a7a1aa6457993f0
/archive/Decathlon-meta-deeplearning/meta.py
a88a9223b3f618d89077beeb4ad102798afaf8d7
[ "MIT" ]
permissive
tueimage/meta-segmentation-msc-2018
0cb9b25c75352ab023dc307f956fc0577fee119b
d497e6ea99b89fdb54e11568452894f022269043
refs/heads/master
2020-04-07T13:16:01.725093
2019-12-13T18:02:00
2019-12-13T18:02:00
158,400,064
3
1
null
null
null
null
UTF-8
Python
false
false
3,253
py
# from meta_network import meta_learner from data import Data, MetaData from utils import subset_index_to_address from utils import meta_pred_generator, historyPlot, create_data_subsets, dice_coef_loss, auc, mean_iou, dice_coef from keras.optimizers import Adam from meta_network import meta_learner from networks import EncoderDecoderNetwork from tqdm import tqdm import os os.environ["CUDA_VISIBLE_DEVICES"] = "1" def main(): tasks_list= ['Task10_Colon','Task01_BrainTumour','Task02_Heart','Task03_Liver','Task04_Hippocampus','Task05_Prostate', 'Task06_Lung', 'Task07_Pancreas', 'Task08_HepaticVessel'] # ,'Task09_Spleen', feature_extractors = ['VGG16', 'ResNet50', 'MobileNetV1'] meta_data = MetaData('t', 's') for task in tasks_list: # try: # data = Data(task) # # data.train_size = 1000 # data.val_size = 100 # data.imageDimensions = (224, 224) # meta_subset_size = 5 # nr_of_meta_subsets = 5 # # data.load_training_data() # data.load_valid_data() # data.get_meta_subsets(nr_of_meta_subsets, meta_subset_size) for fe in feature_extractors: meta_inter = MetaData(task, fe) try: meta_inter.load() for x in range(5): meta_data.total_addresses.append(meta_inter.addresses[x]) meta_data.total_results.append(meta_inter.results[x]) except: print("oei") # meta_data = MetaData(task, fe) # struct = EncoderDecoderNetwork(fe, 2) # struct.task = task # struct.minibatch_size = 5 # struct.epochs = 10 # struct.imageDimensions = (224, 224) # struct.build_encoder() # struct.build_decoder() # struct.load_weights() # struct.model.compile(optimizer = Adam(lr = 1e-5), loss = dice_coef_loss, metrics = ['accuracy', auc, mean_iou, dice_coef]) # for subset in range(data.meta_subsets.shape[0]): # addresses = subset_index_to_address(data.meta_subsets[subset, :], data.train_data) # meta_data.addresses.append(addresses) # result = struct.get_meta_data(addresses)[2:] # print(result) # meta_data.results.append(result) # meta_data.save() print(meta_data.total_addresses) print(meta_data.total_results) print(len(meta_data.total_results)) # except: # print("MASSIVE FAIL") meta_model = meta_learner('VGG16') meta_model.build_feature_extractor() meta_model.build_meta_model() meta_model.train(meta_data.total_addresses, meta_data.total_results, (224, 224), 1) # historyPlot(meta_model.history, "testmeta.png") # meta_model.save_model() # meta_model.load_weights() # for x in range(30): # pred = meta_model.model.predict_generator(meta_pred_generator(meta_data.total_addresses[x], meta_model.minibatch_size, (224,224)), steps = 1) # print("pred: {}".format(pred)) # print("result: {}".format(meta_data.total_results[x])) if __name__ == '__main__': main()
[ "45205052+tjvsonsbeek@users.noreply.github.com" ]
45205052+tjvsonsbeek@users.noreply.github.com
595ee2d2c462822d03e02122bed94ccf834309c7
5ed917ada5766c0a028914f2c15549b6a9095b53
/pyt/bin/jupyter-labextension
03c9c79c4c5f4cbadfdd6106350523ec20a32373
[]
no_license
shreyanshu007/backup
eb53c3cc44b17e1bcaa42ff2f46ea9364f1bcdfc
5a0a4c9715375ae224db8c0f427f847022a9af02
refs/heads/master
2023-01-12T13:33:04.174671
2020-06-12T06:00:53
2020-06-12T06:00:53
201,099,769
0
0
null
2023-01-04T06:33:29
2019-08-07T17:46:24
Python
UTF-8
Python
false
false
250
#!/home/shreyanshu/sem_fir/pyt/bin/python3 # -*- coding: utf-8 -*- import re import sys from jupyterlab.labextensions import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "2016csb1060@iitrpr.ac.in" ]
2016csb1060@iitrpr.ac.in
23159149d8d9130887e04ea1d4f0fd74e9ad9bd7
4e8ae597e7f5010c4dd7208226c3cda99b11e561
/blog/models.py
eee192ab822d24d4e5c69c3761f3737ecc6181a8
[]
no_license
peckzl/django_bloggy
a50999c0f527854a5af5e655155c56c1e8e051b0
cf3e9d3b052f0504256c87c7fab8d1a199fe586a
refs/heads/master
2021-01-10T05:28:32.091423
2015-11-05T00:13:44
2015-11-05T00:13:44
45,576,036
0
0
null
null
null
null
UTF-8
Python
false
false
621
py
from django.db import models from uuslug import uuslug class Post(models.Model): created_at = models.DateTimeField(auto_now_add=True) title = models.CharField(max_length=100) content = models.TextField() tag = models.CharField(max_length=20, blank=True, null=True) image = models.ImageField(upload_to="images", blank=True, null=True) views = models.IntegerField(default=0) slug = models.CharField(max_length=100, unique=True) def __unicode__(self): return self.title def save(self, *args, **kwargs): self.slug = uuslug(self.title, instance=self, max_length=100) super(Post, self).save(*args, **kwargs)
[ "peckzl@clarkson.edu" ]
peckzl@clarkson.edu
91e700c2b68497c13aa063de195ee61e65a3dc8e
2552b11e55ebeab1ceb16f43ac2117e67292616d
/corpus_health/spiders/spider_medlive.py
e8e18783254833aaa1ce4082915b8919205eb8f3
[]
no_license
bzqweiyi/corpus
7271fc93ad0c91dcff90de9c123cd9aa90222ff3
69793804161361bf79944043aa786e96d3c14c9d
refs/heads/master
2020-04-02T02:12:32.808658
2018-10-20T11:12:50
2018-10-20T11:12:50
153,897,293
0
0
null
null
null
null
UTF-8
Python
false
false
4,815
py
#!/usr/bin/env python3.6 # -*- coding: utf-8 -*- """ 中国好中医网:祛湿 James """ from scrapy.spiders import Rule from scrapy.linkextractors import LinkExtractor from corpus_health.items import ArticlespiderItem from scrapy_redis.spiders import RedisCrawlSpider from scrapy_redis.spiders import Spider from scrapy.http import Request from scrapy.http import FormRequest import requests import re from corpus_health.Util.LogHandler import LogHandler logger = LogHandler(__name__, stream=True) class Ask999Spider(RedisCrawlSpider): handle_httpstatus_list = [404, 403, 500] name = 'ymt' allowed_domains = ['www.medlive.cn/'] start_urls = "http://news.medlive.cn/all/info-progress/list.html?ver=branch" # "http://www.zghzyw.com/zyrm/fx/" # 丰胸 # "http://www.zghzyw.com/qushi" # 祛湿 redis_key = 'ymt:start_urls' rules = ( Rule(LinkExtractor(allow=r"http://www.medlive.cn/\d+/$"), callback="parse", follow=False), # Rule(LinkExtractor(allow=()), callback="parse_detail_mongo", follow=False), ) # def start_requests(self): # url = "" # requests = [] # for i in range(0, 100): # formdata = { # "page": str(i), # "submit_type": "ajax", # "ac": "research_branch", # "div_type": "all", # "model_type": "info", # "cat_type": "research"} # request = FormRequest(url, callback=self.parse, formdata=formdata) # requests.append(request) # return requests def parse(self, response): urls = [ "http://news.medlive.cn/infect/info-progress/show-149976_171.html", "http://news.medlive.cn/heart/info-progress/show-149938_129.html"] # f"http://news.medlive.cn/psy/info-progress/show-149946_60.html", # f"http://news.medlive.cn/endocr/info-progress/show-149951_46.html", # f"http://news.medlive.cn/endocr/info-progress/show-149948_46.html", # f"http://news.medlive.cn/imm/info-progress/show-149926_166.html"] for url in urls: print(f"url, {url}") try: # meta = {'dont_redirect': False} yield Request(url, callback=self.parse_detail_mongo, dont_filter=True) except Exception as e: print(e) # def parse_next(self, response): # # nextpath = '//*[@id="main"]/div[1]/div/div[' + str(i) + "]" + "/div[1]/h3/a/@href" # # url = response.xpath(nextpath) # urls = response.xpath('.//div[@class="u-post"]//h3/a/@href').extract() # for url in urls: # newpath = "http://www.zghzyw.com" + url # print(f"newpath,{newpath}") # yield Request(newpath, callback=self.parse_detail_mongo, dont_filter=True) def parse_detail_mongo(self, response): item = ArticlespiderItem() try: # time.sleep(random.uniform(1.1, 5)) # 获取文章url & title item['url'] = response.url print("url: ", item['url']) # response.xpath('//div[@class="u-post"]').extract() try: title = response.xpath('.//div[@id="content"]//div/h1/text()').extract() title = self.filter_tags_blank(title) except Exception as e: title = "" print("title :", e) try: # 获取文章内容" //*[@id="content"]/div/div[3]/text()" content = "".join(response.xpath('.//div[@id="content"]//div/p/span/text()').extract()) content = self.filter_tags_blank(content) # position = response.xpath('//body/div/div/div/div/b/text()').extract()[0] # category0 = response.xpath('//body/div/div/div/div/a/text()').extract()[0] # category1 = response.xpath('//body/div/div/div/div/a/text()').extract()[1] # category = position + ": >" + category0 + ">" + category1 # category = self.filter_tags_blank(title) # print("category :", category) except Exception as e: content = "" print("content :", e) item['title'] = title item['content'] = content print(f"title: ,{title}") print(f"descText, {content}") # item['category'] = category yield item except Exception as e: print("Error2 :", e) logger.info("匹配信息出错。错误原因:") logger.info(e) """ 去掉html标签和空格 """ def filter_tags_blank(self, str): p = re.compile('<[^>]+>').sub("", str) return "".join(p.split())
[ "bzqweiyi@163.com" ]
bzqweiyi@163.com
62ce19f3d0f04ce110c1dd241445d520cdfc6c0c
5e6d8b9989247801718dd1f10009f0f7f54c1eb4
/sdk/python/pulumi_azure_native/containerservice/v20210901/private_endpoint_connection.py
2c0b95aa60f3e4a88ea7d42d3e4f40430d813fb8
[ "BSD-3-Clause", "Apache-2.0" ]
permissive
vivimouret29/pulumi-azure-native
d238a8f91688c9bf09d745a7280b9bf2dd6d44e0
1cbd988bcb2aa75a83e220cb5abeb805d6484fce
refs/heads/master
2023-08-26T05:50:40.560691
2021-10-21T09:25:07
2021-10-21T09:25:07
null
0
0
null
null
null
null
UTF-8
Python
false
false
13,975
py
# coding=utf-8 # *** WARNING: this file was generated by the Pulumi SDK Generator. *** # *** Do not edit by hand unless you're certain you know what you are doing! *** import warnings import pulumi import pulumi.runtime from typing import Any, Mapping, Optional, Sequence, Union, overload from ... import _utilities from . import outputs from ._enums import * from ._inputs import * __all__ = ['PrivateEndpointConnectionArgs', 'PrivateEndpointConnection'] @pulumi.input_type class PrivateEndpointConnectionArgs: def __init__(__self__, *, private_link_service_connection_state: pulumi.Input['PrivateLinkServiceConnectionStateArgs'], resource_group_name: pulumi.Input[str], resource_name: pulumi.Input[str], private_endpoint: Optional[pulumi.Input['PrivateEndpointArgs']] = None, private_endpoint_connection_name: Optional[pulumi.Input[str]] = None): """ The set of arguments for constructing a PrivateEndpointConnection resource. :param pulumi.Input['PrivateLinkServiceConnectionStateArgs'] private_link_service_connection_state: A collection of information about the state of the connection between service consumer and provider. :param pulumi.Input[str] resource_group_name: The name of the resource group. :param pulumi.Input[str] resource_name: The name of the managed cluster resource. :param pulumi.Input['PrivateEndpointArgs'] private_endpoint: The resource of private endpoint. :param pulumi.Input[str] private_endpoint_connection_name: The name of the private endpoint connection. """ pulumi.set(__self__, "private_link_service_connection_state", private_link_service_connection_state) pulumi.set(__self__, "resource_group_name", resource_group_name) pulumi.set(__self__, "resource_name", resource_name) if private_endpoint is not None: pulumi.set(__self__, "private_endpoint", private_endpoint) if private_endpoint_connection_name is not None: pulumi.set(__self__, "private_endpoint_connection_name", private_endpoint_connection_name) @property @pulumi.getter(name="privateLinkServiceConnectionState") def private_link_service_connection_state(self) -> pulumi.Input['PrivateLinkServiceConnectionStateArgs']: """ A collection of information about the state of the connection between service consumer and provider. """ return pulumi.get(self, "private_link_service_connection_state") @private_link_service_connection_state.setter def private_link_service_connection_state(self, value: pulumi.Input['PrivateLinkServiceConnectionStateArgs']): pulumi.set(self, "private_link_service_connection_state", value) @property @pulumi.getter(name="resourceGroupName") def resource_group_name(self) -> pulumi.Input[str]: """ The name of the resource group. """ return pulumi.get(self, "resource_group_name") @resource_group_name.setter def resource_group_name(self, value: pulumi.Input[str]): pulumi.set(self, "resource_group_name", value) @property @pulumi.getter(name="resourceName") def resource_name(self) -> pulumi.Input[str]: """ The name of the managed cluster resource. """ return pulumi.get(self, "resource_name") @resource_name.setter def resource_name(self, value: pulumi.Input[str]): pulumi.set(self, "resource_name", value) @property @pulumi.getter(name="privateEndpoint") def private_endpoint(self) -> Optional[pulumi.Input['PrivateEndpointArgs']]: """ The resource of private endpoint. """ return pulumi.get(self, "private_endpoint") @private_endpoint.setter def private_endpoint(self, value: Optional[pulumi.Input['PrivateEndpointArgs']]): pulumi.set(self, "private_endpoint", value) @property @pulumi.getter(name="privateEndpointConnectionName") def private_endpoint_connection_name(self) -> Optional[pulumi.Input[str]]: """ The name of the private endpoint connection. """ return pulumi.get(self, "private_endpoint_connection_name") @private_endpoint_connection_name.setter def private_endpoint_connection_name(self, value: Optional[pulumi.Input[str]]): pulumi.set(self, "private_endpoint_connection_name", value) class PrivateEndpointConnection(pulumi.CustomResource): @overload def __init__(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, private_endpoint: Optional[pulumi.Input[pulumi.InputType['PrivateEndpointArgs']]] = None, private_endpoint_connection_name: Optional[pulumi.Input[str]] = None, private_link_service_connection_state: Optional[pulumi.Input[pulumi.InputType['PrivateLinkServiceConnectionStateArgs']]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, resource_name_: Optional[pulumi.Input[str]] = None, __props__=None): """ A private endpoint connection :param str resource_name: The name of the resource. :param pulumi.ResourceOptions opts: Options for the resource. :param pulumi.Input[pulumi.InputType['PrivateEndpointArgs']] private_endpoint: The resource of private endpoint. :param pulumi.Input[str] private_endpoint_connection_name: The name of the private endpoint connection. :param pulumi.Input[pulumi.InputType['PrivateLinkServiceConnectionStateArgs']] private_link_service_connection_state: A collection of information about the state of the connection between service consumer and provider. :param pulumi.Input[str] resource_group_name: The name of the resource group. :param pulumi.Input[str] resource_name_: The name of the managed cluster resource. """ ... @overload def __init__(__self__, resource_name: str, args: PrivateEndpointConnectionArgs, opts: Optional[pulumi.ResourceOptions] = None): """ A private endpoint connection :param str resource_name: The name of the resource. :param PrivateEndpointConnectionArgs args: The arguments to use to populate this resource's properties. :param pulumi.ResourceOptions opts: Options for the resource. """ ... def __init__(__self__, resource_name: str, *args, **kwargs): resource_args, opts = _utilities.get_resource_args_opts(PrivateEndpointConnectionArgs, pulumi.ResourceOptions, *args, **kwargs) if resource_args is not None: __self__._internal_init(resource_name, opts, **resource_args.__dict__) else: __self__._internal_init(resource_name, *args, **kwargs) def _internal_init(__self__, resource_name: str, opts: Optional[pulumi.ResourceOptions] = None, private_endpoint: Optional[pulumi.Input[pulumi.InputType['PrivateEndpointArgs']]] = None, private_endpoint_connection_name: Optional[pulumi.Input[str]] = None, private_link_service_connection_state: Optional[pulumi.Input[pulumi.InputType['PrivateLinkServiceConnectionStateArgs']]] = None, resource_group_name: Optional[pulumi.Input[str]] = None, resource_name_: Optional[pulumi.Input[str]] = None, __props__=None): if opts is None: opts = pulumi.ResourceOptions() if not isinstance(opts, pulumi.ResourceOptions): raise TypeError('Expected resource options to be a ResourceOptions instance') if opts.version is None: opts.version = _utilities.get_version() if opts.id is None: if __props__ is not None: raise TypeError('__props__ is only valid when passed in combination with a valid opts.id to get an existing resource') __props__ = PrivateEndpointConnectionArgs.__new__(PrivateEndpointConnectionArgs) __props__.__dict__["private_endpoint"] = private_endpoint __props__.__dict__["private_endpoint_connection_name"] = private_endpoint_connection_name if private_link_service_connection_state is None and not opts.urn: raise TypeError("Missing required property 'private_link_service_connection_state'") __props__.__dict__["private_link_service_connection_state"] = private_link_service_connection_state if resource_group_name is None and not opts.urn: raise TypeError("Missing required property 'resource_group_name'") __props__.__dict__["resource_group_name"] = resource_group_name if resource_name_ is None and not opts.urn: raise TypeError("Missing required property 'resource_name_'") __props__.__dict__["resource_name"] = resource_name_ __props__.__dict__["name"] = None __props__.__dict__["provisioning_state"] = None __props__.__dict__["type"] = None alias_opts = pulumi.ResourceOptions(aliases=[pulumi.Alias(type_="azure-nextgen:containerservice/v20210901:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20200601:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20200601:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20200701:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20200701:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20200901:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20200901:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20201101:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20201101:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20201201:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20201201:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20210201:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20210201:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20210301:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20210301:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20210501:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20210501:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20210701:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20210701:PrivateEndpointConnection"), pulumi.Alias(type_="azure-native:containerservice/v20210801:PrivateEndpointConnection"), pulumi.Alias(type_="azure-nextgen:containerservice/v20210801:PrivateEndpointConnection")]) opts = pulumi.ResourceOptions.merge(opts, alias_opts) super(PrivateEndpointConnection, __self__).__init__( 'azure-native:containerservice/v20210901:PrivateEndpointConnection', resource_name, __props__, opts) @staticmethod def get(resource_name: str, id: pulumi.Input[str], opts: Optional[pulumi.ResourceOptions] = None) -> 'PrivateEndpointConnection': """ Get an existing PrivateEndpointConnection resource's state with the given name, id, and optional extra properties used to qualify the lookup. :param str resource_name: The unique name of the resulting resource. :param pulumi.Input[str] id: The unique provider ID of the resource to lookup. :param pulumi.ResourceOptions opts: Options for the resource. """ opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id)) __props__ = PrivateEndpointConnectionArgs.__new__(PrivateEndpointConnectionArgs) __props__.__dict__["name"] = None __props__.__dict__["private_endpoint"] = None __props__.__dict__["private_link_service_connection_state"] = None __props__.__dict__["provisioning_state"] = None __props__.__dict__["type"] = None return PrivateEndpointConnection(resource_name, opts=opts, __props__=__props__) @property @pulumi.getter def name(self) -> pulumi.Output[str]: """ The name of the private endpoint connection. """ return pulumi.get(self, "name") @property @pulumi.getter(name="privateEndpoint") def private_endpoint(self) -> pulumi.Output[Optional['outputs.PrivateEndpointResponse']]: """ The resource of private endpoint. """ return pulumi.get(self, "private_endpoint") @property @pulumi.getter(name="privateLinkServiceConnectionState") def private_link_service_connection_state(self) -> pulumi.Output['outputs.PrivateLinkServiceConnectionStateResponse']: """ A collection of information about the state of the connection between service consumer and provider. """ return pulumi.get(self, "private_link_service_connection_state") @property @pulumi.getter(name="provisioningState") def provisioning_state(self) -> pulumi.Output[str]: """ The current provisioning state. """ return pulumi.get(self, "provisioning_state") @property @pulumi.getter def type(self) -> pulumi.Output[str]: """ The resource type. """ return pulumi.get(self, "type")
[ "noreply@github.com" ]
noreply@github.com
396d3be1f2a5e259471ee0be5f9b6850177f96e3
b648a0ff402d23a6432643879b0b81ebe0bc9685
/benchmark/tslintbasic/thrift/run.py
6b63c1b1ffc1c7036f1224f0530a63f3d6a08ca5
[ "Apache-2.0" ]
permissive
jviotti/binary-json-size-benchmark
4712faca2724d47d23efef241983ce875dc71cee
165b577884ef366348bf48042fddf54aacfe647a
refs/heads/main
2023-04-18T01:40:26.141995
2022-12-19T13:25:35
2022-12-19T13:25:35
337,583,132
21
1
Apache-2.0
2022-12-17T21:53:56
2021-02-10T01:18:05
C++
UTF-8
Python
false
false
581
py
def encode(json, schema): payload = schema.Main() payload.rules = schema.Rules() payload.rules.orderedImports = schema.OrderedImports() payload.rules.orderedImports.options = schema.Options() payload.rules.orderedImports.options.groupedImports = \ json['rules']['ordered-imports']['options']['grouped-imports'] return payload def decode(payload): return { 'rules': { 'ordered-imports': { 'options': { 'grouped-imports': payload.rules.orderedImports.options.groupedImports } } } }
[ "jv@jviotti.com" ]
jv@jviotti.com
d37a57bf782e04065867f737dfe764d33b9abd38
da54e1256c8e66c78c069331af02a55a00f36faa
/new_python/day2/ch6_1.py
3abcbed2eba1d17b92141354232f795b56841de6
[]
no_license
tghyyhjuujki/Study2
5f7642d5c65176c46cf66b6341a9cdb8cd196400
fffa315ebbe6088cdba219515c8d5ce1b5a4b8a6
refs/heads/master
2023-02-10T15:51:36.757719
2021-01-01T19:34:54
2021-01-01T19:34:54
276,801,852
0
0
null
null
null
null
UTF-8
Python
false
false
82
py
import math #a math.floor(-2.8) #b abs(round(-4.3)) #c math.ceil(math.sin(34.5))
[ "tghyyhju@gmail.com" ]
tghyyhju@gmail.com
9ca07e7be454f472e705453ea355ee2b9fe48e9d
f8cc2cbde9aafca45f1ecd7fe2e0296d536cf94f
/src/ls-files.py
933b6d6ea1a18eaf2f8f7293bfc7369a7ca957bc
[]
no_license
minekoa/mgit
aa6dbc9c971e77ce0f5005e99c4ebd00a5f8feee
de137829be2cd0098309a8dfdab8097f190f9b80
refs/heads/master
2021-01-23T15:41:57.708857
2015-05-13T14:30:31
2015-05-13T14:30:31
35,554,477
1
0
null
null
null
null
UTF-8
Python
false
false
340
py
#!/usr/bin/env python #-*- coding: shift_jis -*- from mgitlib import * import sys import os import os.path import struct if __name__ == '__main__': db = GitDB() index = GitIndex() with db.openIndexFile('rb') as rf: index.unpack(rf) for key, item in index.rows.items(): print item
[ "mineko.orange@gmail.com" ]
mineko.orange@gmail.com
84f9eee0a4ffea06d2accb83ecb72ac8b44d7b69
a1bd103181681e1ea0af8859585e3ce599801335
/musictagger/handlers/numbering.py
130631e398f3a784589e8b35021b0575309f7358
[]
no_license
Tbsc/musictagger
3ce82bf1cf6bd9b265d787014ce23cfb9fdd8103
c01a3ebdb573f6a3ec5957bb876170192ecb73b8
refs/heads/master
2021-12-22T19:18:46.876057
2021-12-18T12:05:54
2021-12-18T12:05:54
99,435,569
0
0
null
null
null
null
UTF-8
Python
false
false
927
py
import re # ensure numbering always consists of at least 2 digits, adding zeros if needed def check(filename): # if numbering is correct, return function with current filename if check_correct(filename): return filename new_filename = filename # begin checks new_filename = check_single_digit(new_filename) return new_filename # checks if numbering is formatted correctly def check_correct(filename): correct_pattern = re.compile("\d\d.*") if correct_pattern.match(filename): print("Numbering is valid") return True return False # check if numbering is only 1 digit def check_single_digit(filename): single_digit_pattern = re.compile("\d.*") if single_digit_pattern.match(filename): # just append a 0 before the number print("Numbering consists of a single digit, prepending a zero") return "0" + filename return filename
[ "danielben60@gmail.com" ]
danielben60@gmail.com
5eb3ad1011c8c6cf69038228cced51111ccd68ab
e6d24866b239362b84d4192ece761c9919f8f323
/HashCode 2018/main.py
386a27c316391568ed76b6639ecdaad6d858ee33
[]
no_license
CodiceLoco/code-challenges
ef6f9d2bd1770dea21d9a419ea10c47f5ff96d86
4fd265d86cdaa05eac756d8a07b02c98fe88e50f
refs/heads/master
2020-03-22T16:04:53.629023
2019-05-08T13:49:32
2019-05-08T13:49:32
140,301,446
1
0
null
null
null
null
UTF-8
Python
false
false
692
py
from solution import try_to_solve, read_file from scoring import JudgeSystem from os import listdir INPUT_FOLDER = 'input' OUTPUT_FOLDER = 'output' FILES = sorted(map(lambda f: f.split('.')[0], listdir(INPUT_FOLDER))) INPUT_FILES = list(map(lambda f: f'{INPUT_FOLDER}/{f}.in', FILES)) OUTPUT_FILES = list(map(lambda f: f'{OUTPUT_FOLDER}/{f}.out', FILES)) total = 0 for input_file, output_file in zip(INPUT_FILES, OUTPUT_FILES): params, ride = read_file(input_file) try_to_solve(params, ride, output_file) j = JudgeSystem(input_file, output_file) score = int(j.score) total += score print(f'Scored {score} points with {input_file}') print(f'Total: {total} points')
[ "marcon.niccolo@gmail.com" ]
marcon.niccolo@gmail.com
5de03e92e379b9d36dae4594c56cd3ed06ad61cc
8a60ed6b07bb361cea238462689f51ff2f220e0a
/Final Project/tags.py
e14052985b3bb540146a1554a48b96f0db7fb37f
[]
no_license
joshsee/P2-Data-Wrangling-with-MongoDB
3b4b9820e7e6d1ac58006570cf2ee3330766d759
658d084bfd7459349018cd1bdea11ef4b3acdd24
refs/heads/master
2016-09-06T00:21:19.961782
2015-03-25T15:34:14
2015-03-25T15:34:14
29,285,468
0
0
null
null
null
null
UTF-8
Python
false
false
2,427
py
#!/usr/bin/env python # -*- coding: utf-8 -*- import xml.etree.ElementTree as ET import pprint import re """ Your task is to explore the data a bit more. Before you process the data and add it into MongoDB, you should check the "k" value for each "<tag>" and see if they can be valid keys in MongoDB, as well as see if there are any other potential problems. We have provided you with 3 regular expressions to check for certain patterns in the tags. As we saw in the quiz earlier, we would like to change the data model and expand the "addr:street" type of keys to a dictionary like this: {"address": {"street": "Some value"}} So, we have to see if we have such tags, and if we have any tags with problematic characters. Please complete the function 'key_type'. """ lower = re.compile(r'^([a-z]|_)*$') lower_colon = re.compile(r'^([a-z]|_)*:([a-z]|_)*$') problemchars = re.compile(r'[=\+/&<>;\'"\?%#$@\,\. \t\r\n]') def key_type(element, keys): if element.tag == "tag": for name, value in element.items(): if name == 'k': if lower.search(value): keys["lower"] += 1 elif lower_colon.search(value): keys["lower_colon"] += 1 elif problemchars.search(value): <<<<<<< HEAD try: print value except UnicodeEncodeError: pass keys["problemchars"] += 1 else: try: print value except UnicodeEncodeError: pass ======= keys["problemchars"] += 1 else: >>>>>>> origin/master keys["other"] += 1 return keys def process_map(filename): keys = {"lower": 0, "lower_colon": 0, "problemchars": 0, "other": 0} for _, element in ET.iterparse(filename): keys = key_type(element, keys) return keys def test(): # You can use another testfile 'map.osm' to look at your solution # Note that the assertions will be incorrect then. <<<<<<< HEAD keys = process_map('hong-kong.osm') # pprint.pprint(keys) ======= keys = process_map('hong-kong_china.osm') pprint.pprint(keys) # assert keys == {'lower': 5, 'lower_colon': 0, 'other': 1, 'problemchars': 1} >>>>>>> origin/master if __name__ == "__main__": test()
[ "ychian@gmail.com" ]
ychian@gmail.com
64e5d54acec6c968bbdc376004abbfcebe376ec6
a4fba2d2aecbf299d7c32741b5572ed8cca6600c
/scrapydeme/car_scrapy/car_scrapy/spiders/autohome_spider.py
23663243541210aee441e3f8eb8f52ace5b9fa98
[]
no_license
rookiexjl/scrapydemo
22a42565b1054d5f06512c7b286bb14858d0c141
dc1a373fb583d659898a50f9e6ddf0dc0bb4c69d
refs/heads/master
2021-01-23T06:01:15.220988
2017-06-12T03:18:13
2017-06-12T03:18:13
93,006,589
1
1
null
null
null
null
UTF-8
Python
false
false
3,020
py
# coding=utf-8 from scrapy import Request from scrapy import Spider from car_scrapy.items import CarScrapyItem class AutohomeSpider(Spider): name = "autohome.new" start_urls = [ "http://www.autohome.com.cn/grade/carhtml/A.html", "http://www.autohome.com.cn/grade/carhtml/B.html", "http://www.autohome.com.cn/grade/carhtml/C.html", "http://www.autohome.com.cn/grade/carhtml/D.html", "http://www.autohome.com.cn/grade/carhtml/E.html", "http://www.autohome.com.cn/grade/carhtml/F.html", "http://www.autohome.com.cn/grade/carhtml/G.html", "http://www.autohome.com.cn/grade/carhtml/H.html", "http://www.autohome.com.cn/grade/carhtml/I.html", "http://www.autohome.com.cn/grade/carhtml/J.html", "http://www.autohome.com.cn/grade/carhtml/K.html", "http://www.autohome.com.cn/grade/carhtml/L.html", "http://www.autohome.com.cn/grade/carhtml/M.html", "http://www.autohome.com.cn/grade/carhtml/N.html", "http://www.autohome.com.cn/grade/carhtml/O.html", "http://www.autohome.com.cn/grade/carhtml/P.html", "http://www.autohome.com.cn/grade/carhtml/Q.html", "http://www.autohome.com.cn/grade/carhtml/R.html", "http://www.autohome.com.cn/grade/carhtml/S.html", "http://www.autohome.com.cn/grade/carhtml/T.html", "http://www.autohome.com.cn/grade/carhtml/U.html", "http://www.autohome.com.cn/grade/carhtml/V.html", "http://www.autohome.com.cn/grade/carhtml/W.html", "http://www.autohome.com.cn/grade/carhtml/X.html", "http://www.autohome.com.cn/grade/carhtml/Y.html", "http://www.autohome.com.cn/grade/carhtml/Z.html", ] def parse(self, response): xPath = u'//ul[@class="rank-list-ul"]/li/h4/a/@href' resList1 = response.xpath(xPath).extract() for i in range(1, len(resList1) + 1): # for i in range(1,2): carURL = resList1[i - 1] #print carURL # baseItem['brand'] = carURL # baseItem['brand'] = carURL meta={'brand': baseItem}, yield Request(carURL, callback=self.parse_detail) def parse_detail(self,response): baseItem = CarScrapyItem() xPath = u'//div[@class="subnav-title-name"]/a/text()' resList1 = response.xpath(xPath).extract() baseItem['brand'] = resList1[0] xPath = u'//div[@class="subnav-title-name"]/a/h1/text()' name = response.xpath(xPath).extract() if len(name)>0: baseItem['name'] = name[0] else: xPath = u'//div[@class="subnav-title-name"]/a/text()' name = response.xpath(xPath).extract() name=name[0].split('-') baseItem['name'] = name[1] xPath = u'//div[@class="subnav-title-name"]/a/@href' url = response.xpath(xPath).extract() URL='m.autohome.com.cn'+url[0] baseItem['url'] = URL print baseItem yield baseItem
[ "18252005734@163.com" ]
18252005734@163.com
2a7751d735e8753be513461b0ed0f324a9cfac53
c067c06fca7a6d371828b75200ed8862dd2977ec
/simple_exercises/profiti/02.av3.py
943796bf35155a2b269172b4af5042e20c864d15
[ "MIT" ]
permissive
ilante/programming_immanuela_englander
1e170ebbadf0de56c35529b0d02268ea96779178
45d51c99b09ae335a67e03ac5ea79fc775bdf0bd
refs/heads/master
2020-11-25T22:54:06.492425
2020-07-14T16:14:27
2020-07-14T16:14:27
228,879,626
0
0
null
null
null
null
UTF-8
Python
false
false
75
py
def av3num(a, b, c): av = (a+b+c)/3 return av print(av3num(1,2,3))
[ "ila@Immanuelas-Air.station" ]
ila@Immanuelas-Air.station
a5dc301a477f897e6ebee27e6add64e1ac8c514c
173ccb51fbf21555a4bc3e7f3f1e28c02a6c3499
/apps/goods/views.py
dcb59e4890bee3e984309bedf473ed8fce8b1e1e
[]
no_license
lingyunds/myproject
ec9bcb9a5917bd73377ecedc112f48cac62fb246
d8bf8c784f434680b9584d84387f4a428b69ae1c
refs/heads/master
2023-04-22T17:45:07.249340
2021-05-04T09:13:34
2021-05-04T09:13:34
363,365,906
0
0
null
null
null
null
UTF-8
Python
false
false
5,484
py
from django.shortcuts import render,redirect from django.urls import reverse from django.views import View from django_redis import get_redis_connection from django.core.cache import cache from django.core.paginator import Paginator from apps.goods.models import GoodsSKU,IndexGoodsBanner,IndexPromotionBanner,GoodsType,IndexTypeGoodsBanner from apps.order.models import OrderGoods # Create your views here. class Index(View): def get(self,request): #获取缓存 context = cache.get('index_data') if context is None: #首页商品种类 types = GoodsType.objects.all() #首页轮播商品 goods_banners = IndexGoodsBanner.objects.all().order_by('index') #首页轮播活动商品 promotion_banners = IndexPromotionBanner.objects.all().order_by('index') #首页分类商品 for type in types: image_banners = IndexTypeGoodsBanner.objects.filter(type=type,display_type=1).order_by('index') title_banners = IndexTypeGoodsBanner.objects.filter(type=type,display_type=0).order_by('index') type.image_banners = image_banners type.title_banners = title_banners context = {'types':types, 'goods_banners':goods_banners, 'promotion_banners':promotion_banners, } #没有缓存则设置缓存 cache.set('index_data',context,3600) #获取购物车条目数 user = request.user cart_count = 0 if user.is_authenticated: conn = get_redis_connection('default') cart_key = 'cart_%d'%user.id cart_count = conn.hlen(cart_key) context.update(cart_count=cart_count) return render(request,'index.html',context) class Detail(View): def get(self,request,sku_id): try: sku = GoodsSKU.objects.get(id=sku_id) except GoodsSKU.DoesNotExist: return redirect(reverse('goods:index')) types = GoodsType.objects.all() spu_skus = GoodsSKU.objects.filter(goods=sku.goods).exclude(id=sku_id) new_skus = GoodsSKU.objects.filter(type=sku.type).order_by('-create_time')[:2] sku_comments = OrderGoods.objects.filter(sku=sku).exclude(comment='') user = request.user sku_count = 0 if user.is_authenticated: conn = get_redis_connection('default') cart_key = 'cart_%d' % user.id cart_count = conn.hlen(cart_key) conn = get_redis_connection('default') history_key = 'history_%d'%user.id conn.lrem(history_key,0,sku_id) conn.lpush(history_key,sku_id) conn.ltrim(history_key,0,4) context = {'sku':sku, 'types':types, 'spu_skus':spu_skus, 'new_skus':new_skus, 'sku_comments':sku_comments, 'cart_count':cart_count } return render(request,'detail.html',context) # /list?type_id=种类id&page=页码&sort=排序方式 # /list/种类id/页码/排序方式 # /list/种类id/页码?sort=排序方式 class List(View): def get(self,request,type_id,page): try: type = GoodsType.objects.get(id=type_id) except GoodsType.DoesNotExist: return redirect(reverse('goods:index')) types = GoodsType.objects.all() #设置排序方式 sort = request.GET.get('sort') if sort == 'price': order_by = 'price' elif sort == 'hot': order_by = '-sales' else: sort = 'default' order_by = '-id' skus = GoodsSKU.objects.filter(type=type).order_by('%s'%order_by) #对内容分页 paginator = Paginator(skus,1) try: page = int(page) except Exception as e: page = 1 if page > paginator.num_pages: page = 1 #获取分页内容 page_skus = paginator.page(page) # 1.总页数小于5页,页面上显示所有页码 # 2.如果当前页是前3页,显示1-5页 # 3.如果当前页是后3页,显示后5页 # 4.其他情况,显示当前页的前2页,当前页,当前页的后2页 num_pages = paginator.num_pages if num_pages < 5: pages = range(1, num_pages + 1) elif page <= 3: pages = range(1, 6) elif num_pages - page <= 2: pages = range(num_pages - 4, num_pages + 1) else: pages = range(page - 2, page + 3) new_skus = GoodsSKU.objects.filter(type=type).order_by('-create_time')[:2] user = request.user cart_count = 0 if user.is_authenticated: conn = get_redis_connection('default') cart_key = 'cart_%d' % user.id cart_count = conn.hlen(cart_key) context = {'type':type, 'types':types, 'page_skus':page_skus, 'new_skus': new_skus, 'cart_count':cart_count, 'sort':sort, 'pages':pages, } return render(request,'list.html',context)
[ "200826704@qq.com" ]
200826704@qq.com
41997f3f3425416215fddea8290ac8569a86bb75
6aec91d3d881ccefd167c55417fae9bb3dd80747
/myutils.py
7ca5a4244d5204e041ea308ef2d425923a44f8e9
[]
no_license
dxcv/spread_arbitrage
8d3be4a3207e0c35eb3232e7cac9f9bd7037eba1
855638515f73d5e56852030639eff2d1f04b36aa
refs/heads/master
2020-05-22T21:42:00.371973
2017-07-22T02:19:12
2017-07-22T02:19:12
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,955
py
# -*- coding:utf-8 -*- import logging import sys import json import time _MaZhaoYang = 0 _YangRui = 1 _NanHuaQiHuo = 2 _Instruments = { "IC": ["IC1707", "IC1708"], "IH": ["IH1707", "IH1708"], "IF": ["IF1707", "IF1708"] } def decode(text): return text.decode('gb2312').encode('utf-8') def get_logger(logger_name, output_file): logger = logging.getLogger(logger_name) # 指定logger输出格式 formatter = logging.Formatter('%(asctime)s [%(levelname)-8s]: %(message)s') # 文件日志 file_handler = logging.FileHandler(output_file) file_handler.setFormatter(formatter) # 可以通过setFormatter指定输出格式 # 控制台日志 console_handler = logging.StreamHandler(sys.stdout) console_handler.formatter = formatter # 也可以直接给formatter赋值 # 为logger添加的日志处理器 logger.addHandler(file_handler) logger.addHandler(console_handler) # 指定日志的最低输出级别,默认为WARN级别 logger.setLevel(logging.INFO) return logger def jsonload(): try: with open('jsonfile.json', 'r') as f: data = json.load(f) return data except: return {} class data_window(): def __init__(self, total_length = 600, second_length = 120): self.length = total_length self.second_length = second_length self.datawindow = [] def add(self, data): if self.datawindow == []: self.datawindow.append(data) else: while len(self.datawindow) > self.length - 1: self.datawindow.pop(0) self.datawindow.append(data) def is_grow(self): if len(self.datawindow) <= self.length - 1: return 0 growth_1min = self.datawindow[-1] - self.datawindow[-120] growth_5min = self.datawindow[-1] - self.datawindow[0] if growth_1min > 0 and growth_5min > 0: return 1 elif growth_1min < 0 and growth_5min < 0: return -1 else: return 0 def clear(self): self.datawindow = [] def get_account(account): """ :param account: 2: 南华期货 1:杨睿_SimNow 0:马朝阳_SimNow ... :return: """ if account == 2: # 南华期货 brokerID = '1008' userID = '90095502' password = '222832' mdapi_front = 'tcp://115.238.106.253:41213' trade_front = 'tcp://115.238.106.253:41205' else: # simnow brokerID = b'9999' mdapi_front = b'tcp://180.168.146.187:10010' trade_front = b'tcp://180.168.146.187:10000' if account == 1: userID = b'097138' password = b'285135278' else: userID = b'092120' password = b'mzy187623' return brokerID, userID, password, mdapi_front, trade_front if __name__ == '__main__': from unittest import TestCase
[ "595403043@qq.com" ]
595403043@qq.com
6343515d83aa5cbb5afcc7c49c954ed62aa5cfbb
ef147caaf28199b7961ec40320df4ff080ff3146
/tsdownloaderv1.3.py
364353d72db960c53b5a5896b25f697354c677d3
[]
no_license
DicksonC96/TS-video-downloader
6891868ee42476d99413dd8cb4a563ab922512de
6917f7a32b5c9e0cbef8d587e24e075f940b20ed
refs/heads/main
2023-05-12T14:02:00.242037
2021-06-02T07:24:09
2021-06-02T07:24:09
362,239,382
0
0
null
null
null
null
UTF-8
Python
false
false
1,199
py
import requests import sys def main(url, fname, mini=0, maxi=5000): with open(fname+".ts", 'wb') as f: if not mini==0: for i in range(mini): r = requests.get(url+str(i)+".ts", stream=True) sys.stdout.write("Downloading "+str(i)+".ts ...\n") sys.stdout.flush() for chunk in r.iter_content(chunk_size=None): if chunk: f.write(chunk) r.close() for i in range(mini, maxi): r = requests.get(url+str(i)+".ts", stream=True) if len(r.content) < 179: print("Download finished with "+str(i+1)+" iterations.") break else: sys.stdout.write("Downloading "+str(i)+".ts ...\n") sys.stdout.flush() for chunk in r.iter_content(chunk_size=None): if chunk: f.write(chunk) r.close() url = "https://abcd.voxzer.org/stream/608bd4cc0b8bd18237c8fc6d/1080/index" filename = "Tom Clancy's Without Remorse" ### main(url, filename, minimum_iteration, maximum_iteration) main(url, filename, 1300, 1400)
[ "66625723+Dickson96@users.noreply.github.com" ]
66625723+Dickson96@users.noreply.github.com
df0b59323ca9a839dcf6b4c11f1da303ae237fac
ecd2aa3d12a5375498c88cfaf540e6e601b613b3
/Facebook/Pro105. Construct Binary Tree from Preorder and Inorder Traversal.py
a39da533bff18e1cca864459d11a600e0252ce83
[]
no_license
YoyinZyc/Leetcode_Python
abd5d90f874af5cd05dbed87f76885a1ca480173
9eb44afa4233fdedc2e5c72be0fdf54b25d1c45c
refs/heads/master
2021-09-05T17:08:31.937689
2018-01-29T21:57:44
2018-01-29T21:57:44
103,157,916
0
0
null
null
null
null
UTF-8
Python
false
false
642
py
# Definition for a binary tree node. # class TreeNode: # def __init__(self, x): # self.val = x # self.left = None # self.right = None class Solution: def buildTree(self, preorder, inorder): """ :type preorder: List[int] :type inorder: List[int] :rtype: TreeNode """ if not preorder: return None root = TreeNode(preorder[0]) index = inorder.index(preorder[0]) root.left = self.buildTree(preorder[1:index + 1], inorder[:index]) root.right = self.buildTree(preorder[index + 1:], inorder[index + 1:]) return root
[ "yoyinzyc@gmail.com" ]
yoyinzyc@gmail.com
7fed42466b7f96d16dc21e46e0c0f0514aeb6977
c199f949c04933205a25c71706c80a12d7ba20a6
/webapp/models.py
4b42c5d5b5c7dca25cc112842045a6af2d2bf9ac
[]
no_license
NepaliUtsab/NepaliUtsab.github.io
ea1909017782ba4da794b85b19ba41a9e41aa0ae
7bd91c9cffc116e3daa11d77aa8b4276ed9e77e0
refs/heads/master
2020-03-23T02:17:01.233240
2018-07-14T17:55:16
2018-07-14T17:55:16
140,605,326
0
0
null
null
null
null
UTF-8
Python
false
false
283
py
from django.db import models # Create your models here. class Student(models.Model): firstName = models.CharField(max_length = 10) lastName = models.CharField(max_length = 10) stdId = models.IntegerField() def __str__(self): return self.firstName
[ "utsab@Utsab.local" ]
utsab@Utsab.local
f767561f0b3d9e8b78a4864603261931f59d4217
012c9396f6fd0c88769165b484f77c346b96cb56
/software/artyS7/cmdClass.py
be3bc532207ec15d1bda90d8a099b5f7857efe07
[]
no_license
atfienberg/artyS7
d770209dfddc9860a57626cc8b318b4e7aba3c7c
cd29d49f6d3811e425098715f969853cfdb4872c
refs/heads/master
2021-05-25T21:16:48.620081
2020-11-16T16:43:50
2020-11-16T16:43:50
253,920,188
0
0
null
null
null
null
UTF-8
Python
false
false
6,533
py
################################################################ # Tyler Anderson Thu 07/25/2019_ 9:24:16.32 # # A python script for parsing commands. ################################################################ import logging import string import binascii import sys import struct import crcmod class cmdClass: """A class for handling the command protocol.""" ########################################################### # Data members # Constants S_ACT_DICT = { "swr": 0x0001, # single write "srd": 0x0002, # single read "bwr": 0x8001, # burst write "brd": 0x8002, } # burst read ########################################################### # Methods def __init__(self, adr_dict=None): # Command struct self.cmd = { "ok": True, "raw_cmd": "\x8f\xc7\x00\x02\x0f\xff", # Raw command "raw_rsp": "", # Raw response "hdr": 0x8FC7, # header word "act": 0x0002, # swr, srd, bwr, brd "len": 0x0001, # packet length for burst read/write "adr": 0x0FFF, # a look up for the address "data": 0x0000, # data "crc16": 0x0000, } # CRC16 self.S_ADR_DICT = adr_dict def srd_raw_cmd(self): self.cmd["raw_cmd"] = struct.pack(">H", self.cmd["hdr"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["act"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["adr"]) # print binascii.hexlify(self.cmd['raw_cmd']) def srd_raw_rsp(self): x = int(binascii.hexlify(self.cmd["raw_rsp"]), 16) self.cmd["data"] = x >> 16 self.cmd["crc16"] = x & 0xFFFF # print '%x' % self.cmd['data'] # print '%x' % self.cmd['crc16'] def single_crc16_calc(self): crc16 = crcmod.mkCrcFun(0x18005, rev=False, initCrc=0xFFFF, xorOut=0x0000) # print self.cmd['adr'] # print type(self.cmd['adr']) xstr = hex(self.cmd["adr"])[2:].zfill(4) xstr = xstr + hex(self.cmd["data"])[2:].zfill(4) # print(xstr) checksum16 = int(hex(crc16(bytearray.fromhex(xstr)))[2:], 16) return checksum16 def swr_raw_cmd(self): self.cmd["raw_cmd"] = struct.pack(">H", self.cmd["hdr"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["act"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["adr"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["data"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["crc16"]) # print binascii.hexlify(self.cmd['raw_cmd']) def single_crc16_check(self): checksum16 = self.single_crc16_calc() if self.cmd["crc16"] == checksum16: self.cmd["ok"] = True else: self.cmd["ok"] = False def gse_cmd_str(self): return ( "cmd: act = " + hex(self.cmd["act"]) + ", len = " + hex(self.cmd["len"]) + ", adr = " + hex(self.cmd["adr"]) + ", data = " + hex(self.cmd["data"]) + ", crc16 = " + hex(self.cmd["crc16"]) + ", ok = " + str(self.cmd["ok"]) ) def parse_cmd( self, logging, s_act, h_act, act, s_adr, h_adr, adr, h_length, length, h_data, data, ): # Decode the data if data != None: self.cmd["data"] = data elif h_data != None: self.cmd["data"] = int(h_data, 16) self.cmd["inp_h_data"] = h_data # print 'data is %d' % self.cmd['data'] # Decode the length if length != None: self.cmd["len"] = length elif h_length != None: self.cmd["len"] = int(h_length, 16) # Decode the address if adr != None: self.cmd["adr"] = adr elif h_adr != None: self.cmd["adr"] = int(h_adr, 16) elif s_adr != None: self.cmd["adr"] = self.S_ADR_DICT[s_adr] else: logging.warning( "comClass: ERROR: Must specify one of s_adr, h_adr, or adr! Exiting!" ) exit(-1) # Decode the action if act != None: self.cmd["act"] = act elif h_act != None: self.cmd["act"] = int(h_act, 16) elif s_act != None: self.cmd["act"] = self.S_ACT_DICT[s_act] else: logging.warning( "comClass: ERROR: Must specify either s_act or act! Exiting!" ) exit(-1) # ATF -- burst read and write def bwr_raw_cmd(self): self.cmd["raw_cmd"] = struct.pack(">H", self.cmd["hdr"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["act"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["len"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["adr"]) # pack in the data for word in self.cmd["words_array"]: self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", int(word, 16)) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["crc16"]) def brd_raw_cmd(self): self.cmd["raw_cmd"] = struct.pack(">H", self.cmd["hdr"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["act"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["len"]) self.cmd["raw_cmd"] = self.cmd["raw_cmd"] + struct.pack(">H", self.cmd["adr"]) def brd_raw_rsp(self): # srd_raw_rsp works for for now, # but if the first register is all 0's, they will not be printed self.srd_raw_rsp() def burst_crc16_calc(self): crc16 = crcmod.mkCrcFun(0x18005, rev=False, initCrc=0xFFFF, xorOut=0x0000) xstr = hex(self.cmd["adr"])[2:].zfill(4) xstr = xstr + hex(self.cmd["data"])[2:].rstrip("L").zfill(4 * self.cmd["len"]) checksum16 = int(hex(crc16(bytearray.fromhex(xstr)))[2:], 16) return checksum16 def burst_crc16_check(self): checksum16 = self.burst_crc16_calc() if self.cmd["crc16"] == checksum16: self.cmd["ok"] = True else: self.cmd["ok"] = False
[ "atfienberg@psu.edu" ]
atfienberg@psu.edu
c30096f5fdbe8afa9c1d25074d1d0eebbdc72540
c8414c48d7e2b21674be71aa2596d0df547d090a
/config.py
916379b9a47b5023b6913bb77c385c0bd4d1e18d
[]
no_license
Tyux/PRA_Detection_TF
8cd9882a42c7f02220c9812948f6df2fbd17ee4a
35157a915cca446796ba0d3971e2b162e03c10a9
refs/heads/master
2022-08-02T06:11:46.876012
2020-05-30T21:35:18
2020-05-30T21:35:18
267,481,417
0
0
null
null
null
null
UTF-8
Python
false
false
13,187
py
# -*- coding: utf-8 -*- # File: config.py import numpy as np import os import pprint import six from tensorpack.utils import logger from tensorpack.utils.gpu import get_num_gpu __all__ = ['config', 'finalize_configs'] class AttrDict(): _freezed = False """ Avoid accidental creation of new hierarchies. """ def __getattr__(self, name): if self._freezed: raise AttributeError(name) if name.startswith('_'): # Do not mess with internals. Otherwise copy/pickle will fail raise AttributeError(name) ret = AttrDict() setattr(self, name, ret) return ret def __setattr__(self, name, value): if self._freezed and name not in self.__dict__: raise AttributeError( "Config was freezed! Unknown config: {}".format(name)) super().__setattr__(name, value) def __str__(self): return pprint.pformat(self.to_dict(), indent=1, width=100, compact=True) __repr__ = __str__ def to_dict(self): """Convert to a nested dict. """ return {k: v.to_dict() if isinstance(v, AttrDict) else v for k, v in self.__dict__.items() if not k.startswith('_')} def update_args(self, args): """Update from command line args. """ for cfg in args: keys, v = cfg.split('=', maxsplit=1) keylist = keys.split('.') dic = self for i, k in enumerate(keylist[:-1]): assert k in dir(dic), "Unknown config key: {}".format(keys) dic = getattr(dic, k) key = keylist[-1] oldv = getattr(dic, key) if not isinstance(oldv, str): v = eval(v) setattr(dic, key, v) def freeze(self, freezed=True): self._freezed = freezed for v in self.__dict__.values(): if isinstance(v, AttrDict): v.freeze(freezed) # avoid silent bugs def __eq__(self, _): raise NotImplementedError() def __ne__(self, _): raise NotImplementedError() config = AttrDict() _C = config # short alias to avoid coding # mode flags --------------------- _C.TRAINER = 'horovod' # options: 'horovod', 'replicated' _C.MODE_MASK = False # Faster R-CNN or Mask R-CNN _C.MODE_FPN = True # dataset ----------------------- _C.DATA.BASEDIR = '/work/DeepLearning/Data/objects365/' # All available dataset names are defined in `dataset/coco.py:register_coco`. # All TRAIN dataset will be concatenated for training. _C.DATA.TRAIN = ('objects365_train.json',) # _C.DATA.TRAIN = ('coco_train2014', 'coco_valminusminival2014') # i.e. trainval35k # Each VAL dataset will be evaluated separately (instead of concatenated) _C.DATA.VAL = ('objects365_val.json',) # _C.DATA.VAL = ('coco_minival2014',) # AKA minival2014 # _C.DATA.TEST = ('objects365_Tiny_Testset_images_list.json',) _C.DATA.TEST = ('test',) # These two configs will be populated later inside `finalize_configs`. _C.DATA.NUM_CATEGORY = -1 # without the background class (e.g., 80 for COCO) _C.DATA.CLASS_NAMES = [] # NUM_CLASS (NUM_CATEGORY+1) strings, the first is "BG". # whether the coordinates in the annotations are absolute pixel values, or a relative value in [0, 1] _C.DATA.ABSOLUTE_COORD = True # Number of data loading workers. # In case of horovod training, this is the number of workers per-GPU (so you may want to use a smaller number). # Set to 0 to disable parallel data loading _C.DATA.NUM_WORKERS = 1 # backbone ---------------------- _C.BACKBONE.WEIGHTS = './weights/ImageNet-R50-AlignPadding.npz' # To train from scratch, set it to empty, and set FREEZE_AT to 0 # To train from ImageNet pre-trained models, use the one that matches your # architecture from http://models.tensorpack.com under the 'FasterRCNN' section. # To train from an existing COCO model, use the path to that file, and change # the other configurations according to that model. _C.BACKBONE.RESNET_NUM_BLOCKS = [3, 4, 6, 3] # for resnet50 # _C.BACKBONE.RESNET_NUM_BLOCKS = [2, 2, 2, 2] # RESNET_NUM_BLOCKS = [3, 4, 23, 3] # for resnet101 _C.BACKBONE.FREEZE_AFFINE = False # do not train affine parameters inside norm layers _C.BACKBONE.NORM = 'FreezeBN' # options: FreezeBN, SyncBN, GN, None _C.BACKBONE.FREEZE_AT = 2 # options: 0, 1, 2. How many stages in backbone to freeze (not training) # Use a base model with TF-preferred padding mode, # which may pad more pixels on right/bottom than top/left. # See https://github.com/tensorflow/tensorflow/issues/18213 # In tensorpack model zoo, ResNet models with TF_PAD_MODE=False are marked with "-AlignPadding". # All other models under `ResNet/` in the model zoo are using TF_PAD_MODE=True. # Using either one should probably give the same performance. # We use the "AlignPadding" one just to be consistent with caffe2. _C.BACKBONE.TF_PAD_MODE = False _C.BACKBONE.STRIDE_1X1 = False # True for MSRA models # schedule ----------------------- _C.TRAIN.NUM_GPUS = None # by default, will be set from code _C.TRAIN.WEIGHT_DECAY = 1e-4 _C.TRAIN.BASE_LR = 1e-2 # defined for total batch size=8. Otherwise it will be adjusted automatically _C.TRAIN.WARMUP = 1000 # in terms of iterations. This is not affected by #GPUs _C.TRAIN.WARMUP_INIT_LR = 1e-2 * 0.33 # defined for total batch size=8. Otherwise it will be adjusted automatically _C.TRAIN.STEPS_PER_EPOCH = 500 _C.TRAIN.STARTING_EPOCH = 1 # the first epoch to start with, useful to continue a training # LR_SCHEDULE means equivalent steps when the total batch size is 8. # When the total bs!=8, the actual iterations to decrease learning rate, and # the base learning rate are computed from BASE_LR and LR_SCHEDULE. # Therefore, there is *no need* to modify the config if you only change the number of GPUs. _C.TRAIN.LR_SCHEDULE = [720000, 960000, 1080000] # "1x" schedule in detectron _C.TRAIN.EVAL_PERIOD = 300 # period (epochs) to run evaluation _C.TRAIN.CHECKPOINT_PERIOD = 5 # period (epochs) to save model # preprocessing -------------------- # Alternative old (worse & faster) setting: 600 _C.PREPROC.TRAIN_SHORT_EDGE_SIZE = [800, 800] # [min, max] to sample from _C.PREPROC.TEST_SHORT_EDGE_SIZE = 800 _C.PREPROC.MAX_SIZE = 1333 # mean and std in RGB order. # Un-scaled version: [0.485, 0.456, 0.406], [0.229, 0.224, 0.225] _C.PREPROC.PIXEL_MEAN = [123.675, 116.28, 103.53] _C.PREPROC.PIXEL_STD = [58.395, 57.12, 57.375] # anchors ------------------------- _C.RPN.ANCHOR_STRIDE = 16 _C.RPN.ANCHOR_SIZES = (32, 64, 128, 256, 512) # sqrtarea of the anchor box _C.RPN.ANCHOR_RATIOS = (0.5, 1., 2.) _C.RPN.POSITIVE_ANCHOR_THRESH = 0.7 _C.RPN.NEGATIVE_ANCHOR_THRESH = 0.3 # rpn training ------------------------- _C.RPN.FG_RATIO = 0.5 # fg ratio among selected RPN anchors _C.RPN.BATCH_PER_IM = 256 # total (across FPN levels) number of anchors that are marked valid _C.RPN.MIN_SIZE = 0 _C.RPN.PROPOSAL_NMS_THRESH = 0.7 # Anchors which overlap with a crowd box (IOA larger than threshold) will be ignored. # Setting this to a value larger than 1.0 will disable the feature. # It is disabled by default because Detectron does not do this. _C.RPN.CROWD_OVERLAP_THRESH = 9.99 _C.RPN.HEAD_DIM = 1024 # used in C4 only # RPN proposal selection ------------------------------- # for C4 _C.RPN.TRAIN_PRE_NMS_TOPK = 12000 _C.RPN.TRAIN_POST_NMS_TOPK = 2000 _C.RPN.TEST_PRE_NMS_TOPK = 6000 _C.RPN.TEST_POST_NMS_TOPK = 1000 # if you encounter OOM in inference, set this to a smaller number # for FPN, #proposals per-level and #proposals after merging are (for now) the same # if FPN.PROPOSAL_MODE = 'Joint', these options have no effect _C.RPN.TRAIN_PER_LEVEL_NMS_TOPK = 2000 _C.RPN.TEST_PER_LEVEL_NMS_TOPK = 1000 # fastrcnn training --------------------- _C.FRCNN.BATCH_PER_IM = 512 _C.FRCNN.BBOX_REG_WEIGHTS = [10., 10., 5., 5.] # Slightly better setting: 20, 20, 10, 10 _C.FRCNN.FG_THRESH = 0.5 _C.FRCNN.FG_RATIO = 0.25 # fg ratio in a ROI batch # FPN ------------------------- _C.FPN.ANCHOR_STRIDES = (4, 8, 16, 32, 64) # strides for each FPN level. Must be the same length as ANCHOR_SIZES _C.FPN.PROPOSAL_MODE = 'Level' # 'Level', 'Joint' _C.FPN.NUM_CHANNEL = 256 _C.FPN.NORM = 'None' # 'None', 'GN' # The head option is only used in FPN. For C4 models, the head is C5 _C.FPN.FRCNN_HEAD_FUNC = 'fastrcnn_2fc_head' # choices: fastrcnn_2fc_head, fastrcnn_4conv1fc_{,gn_}head _C.FPN.FRCNN_CONV_HEAD_DIM = 256 _C.FPN.FRCNN_FC_HEAD_DIM = 1024 _C.FPN.MRCNN_HEAD_FUNC = 'maskrcnn_up4conv_head' # choices: maskrcnn_up4conv_{,gn_}head # Mask R-CNN _C.MRCNN.HEAD_DIM = 256 _C.MRCNN.ACCURATE_PASTE = True # slightly more aligned results, but very slow on numpy # Cascade R-CNN, only available in FPN mode _C.FPN.CASCADE = False _C.CASCADE.IOUS = [0.5, 0.6, 0.7] _C.CASCADE.BBOX_REG_WEIGHTS = [[10., 10., 5., 5.], [20., 20., 10., 10.], [30., 30., 15., 15.]] # testing ----------------------- _C.TEST.FRCNN_NMS_THRESH = 0.5 # Smaller threshold value gives significantly better mAP. But we use 0.05 for consistency with Detectron. # mAP with 1e-4 threshold can be found at https://github.com/tensorpack/tensorpack/commit/26321ae58120af2568bdbf2269f32aa708d425a8#diff-61085c48abee915b584027e1085e1043 # noqa _C.TEST.RESULT_SCORE_THRESH = 0.05 _C.TEST.RESULT_SCORE_THRESH_VIS = 0.5 # only visualize confident results _C.TEST.RESULTS_PER_IM = 100 _C.freeze() # avoid typo / wrong config keys def finalize_configs(is_training): """ Run some sanity checks, and populate some configs from others """ _C.freeze(False) # populate new keys now if isinstance(_C.DATA.VAL, six.string_types): # support single string (the typical case) as well _C.DATA.VAL = (_C.DATA.VAL, ) if isinstance(_C.DATA.TRAIN, six.string_types): # support single string _C.DATA.TRAIN = (_C.DATA.TRAIN, ) # finalize dataset definitions ... from dataset import DatasetRegistry datasets = list(_C.DATA.TRAIN) + list(_C.DATA.VAL) _C.DATA.CLASS_NAMES = DatasetRegistry.get_metadata(datasets[0], "class_names") _C.DATA.NUM_CATEGORY = len(_C.DATA.CLASS_NAMES) - 1 assert _C.BACKBONE.NORM in ['FreezeBN', 'SyncBN', 'GN', 'None'], _C.BACKBONE.NORM if _C.BACKBONE.NORM != 'FreezeBN': assert not _C.BACKBONE.FREEZE_AFFINE assert _C.BACKBONE.FREEZE_AT in [0, 1, 2] _C.RPN.NUM_ANCHOR = len(_C.RPN.ANCHOR_SIZES) * len(_C.RPN.ANCHOR_RATIOS) assert len(_C.FPN.ANCHOR_STRIDES) == len(_C.RPN.ANCHOR_SIZES) # image size into the backbone has to be multiple of this number _C.FPN.RESOLUTION_REQUIREMENT = _C.FPN.ANCHOR_STRIDES[3] # [3] because we build FPN with features r2,r3,r4,r5 if _C.MODE_FPN: size_mult = _C.FPN.RESOLUTION_REQUIREMENT * 1. _C.PREPROC.MAX_SIZE = np.ceil(_C.PREPROC.MAX_SIZE / size_mult) * size_mult assert _C.FPN.PROPOSAL_MODE in ['Level', 'Joint'] assert _C.FPN.FRCNN_HEAD_FUNC.endswith('_head') assert _C.FPN.MRCNN_HEAD_FUNC.endswith('_head') assert _C.FPN.NORM in ['None', 'GN'] if _C.FPN.CASCADE: # the first threshold is the proposal sampling threshold assert _C.CASCADE.IOUS[0] == _C.FRCNN.FG_THRESH assert len(_C.CASCADE.BBOX_REG_WEIGHTS) == len(_C.CASCADE.IOUS) if is_training: train_scales = _C.PREPROC.TRAIN_SHORT_EDGE_SIZE if isinstance(train_scales, (list, tuple)) and train_scales[1] - train_scales[0] > 100: # don't autotune if augmentation is on os.environ['TF_CUDNN_USE_AUTOTUNE'] = '1' os.environ['TF_AUTOTUNE_THRESHOLD'] = '0' assert _C.TRAINER in ['horovod', 'replicated'], _C.TRAINER lr = _C.TRAIN.LR_SCHEDULE if isinstance(lr, six.string_types): if lr.endswith("x"): LR_SCHEDULE_KITER = { "{}x".format(k): [1080 * k - 720, 1080 * k - 240, 1080 * k] for k in range(2, 10)} LR_SCHEDULE_KITER["1x"] = [720, 960, 1080] _C.TRAIN.LR_SCHEDULE = [x * 1000 for x in LR_SCHEDULE_KITER[lr]] else: _C.TRAIN.LR_SCHEDULE = eval(lr) # setup NUM_GPUS if _C.TRAINER == 'horovod': import horovod.tensorflow as hvd ngpu = hvd.size() logger.info("Horovod Rank={}, Size={}, LocalRank={}".format( hvd.rank(), hvd.size(), hvd.local_rank())) else: assert 'OMPI_COMM_WORLD_SIZE' not in os.environ ngpu = get_num_gpu() assert ngpu > 0, "Has to train with GPU!" assert ngpu % 8 == 0 or 8 % ngpu == 0, "Can only train with 1,2,4 or >=8 GPUs, but found {} GPUs".format(ngpu) else: # autotune is too slow for inference os.environ['TF_CUDNN_USE_AUTOTUNE'] = '0' ngpu = get_num_gpu() if _C.TRAIN.NUM_GPUS is None: _C.TRAIN.NUM_GPUS = ngpu else: if _C.TRAINER == 'horovod': assert _C.TRAIN.NUM_GPUS == ngpu else: assert _C.TRAIN.NUM_GPUS <= ngpu _C.freeze() logger.info("Config: ------------------------------------------\n" + str(_C))
[ "tong63377@163.com" ]
tong63377@163.com
c96c01d4f6fed322dc79eed64cbca2c8795827f6
d68eec2b77e934022695f8c396c2a597aa1c36be
/REST_API_v2/Users/create_user_contact_method.py
9f5acf27bf4dd7199325b57e5ed4581d3a72fca9
[]
no_license
drummerweed/API_Python_Examples
088570dbef5a7ba6c053cdf4081430d2d7d55417
641784c53628daeceb6f6735ad0de576413c35d8
refs/heads/master
2020-04-04T21:13:43.391354
2018-08-22T17:37:53
2018-08-22T17:37:53
156,279,682
1
0
null
2018-11-05T20:37:49
2018-11-05T20:37:49
null
UTF-8
Python
false
false
2,653
py
#!/usr/bin/env python # # Copyright (c) 2016, PagerDuty, Inc. <info@pagerduty.com> # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of PagerDuty Inc nor the # names of its contributors may be used to endorse or promote products # derived from this software without specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL PAGERDUTY INC BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. import requests import json # Update to match your API key API_KEY = '3c3gRvzx7uGfMYEnWKvF' # Update to match ID of user you want to update ID = 'P0H7Y7J' # Update to match your chosen parameters TYPE = 'email_contact_method' # Can be one of email_contact_method, sms_contact_method, phone_contact_method, or push_notification_contact_method # NOQA ADDRESS = 'insert_email@here.com' LABEL = 'Work' def create_user_contact_method(): url = 'https://api.pagerduty.com/users/{id}/contact_methods'.format(id=ID) headers = { 'Accept': 'application/vnd.pagerduty+json;version=2', 'Authorization': 'Token token={token}'.format(token=API_KEY), 'Content-type': 'application/json' } payload = { 'contact_method': { 'type': TYPE, 'address': ADDRESS, 'label': LABEL } } r = requests.post(url, headers=headers, data=json.dumps(payload)) print 'Status Code: {code}'.format(code=r.status_code) print r.json() if __name__ == '__main__': create_user_contact_method()
[ "lucasfepp@gmail.com" ]
lucasfepp@gmail.com
1c732829737a797fd122d90ea40e90fd3a121ffd
3700ac98bb8f6faf5dbcf20be2f7f5270b3429b9
/server/test/helpers.py
a2da1815e88500c1751a8596d22b323a721663bb
[ "Apache-2.0" ]
permissive
NWCalvank/react-python-starter
47478df0e611a04f33d08ec64e623b3bdf9f7af1
8bee6129f425d6284aba0a9bf1ccce7b696b837c
refs/heads/master
2020-09-27T04:24:40.691221
2019-12-27T19:33:33
2019-12-27T19:33:33
226,389,820
0
0
Apache-2.0
2019-12-06T21:33:35
2019-12-06T18:47:41
Python
UTF-8
Python
false
false
282
py
from app import db from app.api.models.foo import Foo # Helper function to add a sample string to Foo def create_foo_string(foo_string): foo_table_element = Foo(string_field=foo_string) db.session.add(foo_table_element) db.session.commit() return foo_table_element
[ "noreply@github.com" ]
noreply@github.com
ad09f1178489249f0bbe1ac89b8c04992d14a89e
11d0cc8470722b737ac1d83610f227ffb98ea71a
/convert/ahf2csv.py
806bfe4ea9ced908b9a5b9164efc7b63a5b04e65
[]
no_license
EdoardoCarlesi/PyRCODIO
d86e5298877f08b597f92a07aac0c9b634dbfa39
d3f84efb7aeec0032ef6bde839fe9440ee82b392
refs/heads/master
2021-04-03T04:58:26.219884
2021-03-21T12:22:19
2021-03-21T12:22:19
125,036,704
148
0
null
null
null
null
UTF-8
Python
false
false
452
py
''' Python Routines for COsmology and Data I/ (PyRCODIO) v0.2 Edoardo Carlesi 2020 ecarlesi83@gmail.com ahf2csv.py: convert (and compress) AHF halo catalogs to csv files ''' import pandas as pd import sys sys.path.insert(1, '/home/edoardo/CLUES/PyRCODIO/') import read_files as rf this_ahf = sys.argv[1] mpi = sys.argv[2] out_file = this_ahf + '.csv' halo_df = rf.read_ahf_halo(this_ahf, file_mpi=mpi) halo_df.to_csv(out_file)
[ "gatto@nanowar.it" ]
gatto@nanowar.it
3d6a8d3932e7842aed3095ee73c35260a9beb8c0
40021c7443c3b512fd77df1a77855a10c0afd6df
/src/controller/command/track.py
b9796406b68396d87a5f02e48dc2e35b43675ec0
[]
no_license
ginsm/punch
9062cc6a18352a9a6569d2cdd3383563f012bf75
be35587fffdb036f823b13e8df06d9509a348120
refs/heads/master
2021-04-10T01:39:04.296020
2020-03-24T23:59:08
2020-03-24T23:59:08
248,900,501
0
0
null
null
null
null
UTF-8
Python
false
false
681
py
# SECTION IMPORTS - External and internal imports. # ===================================================== # External import re # Internal import view.commands.track as view import model.db as db # SECTION HANDLER - Job selection handlers. # ===================================================== def handler(command, argument): if argument is None: return view.jobNameRequired() if re.search('\/', argument) is not None: return view.invalidCharacter(argument) db.set_state({'job': argument}) if not db.exists(argument): schema = db.get_state('schema') schema['name'] = argument db.write(schema, argument) return view.newSelectedJob(argument)
[ "matt@mgin.me" ]
matt@mgin.me
e128dcb7bacbd5c878d4d8dc67752ce9b8a0b7d1
dbd7be725a36b22a99f3785bcdf2e3866b6e3cc7
/tests/base.py
a2813c552c6c3d551bc6e9e5c9deea5c949d9492
[]
no_license
MatthewMcGonagle/TSP_PictureMaker
9aed2ddebfc8c36c5a835cd4fb3fd75a1650516c
b6df1b6e5ade27750e2974de1c46e46988fb16e9
refs/heads/master
2020-03-19T12:43:52.242516
2019-03-21T19:51:08
2019-03-21T19:51:08
136,536,415
0
0
null
null
null
null
UTF-8
Python
false
false
2,405
py
import unittest import numpy as np import sys import fake_random sys.path.append('..') import tsp_draw.base class TestAnnealerMethods(unittest.TestCase): def __init__(self, *args, **kwargs): unittest.TestCase.__init__(self, *args, **kwargs) angles = np.linspace(0, 2 * np.pi, 8)[:-1] self.vertices = [[np.cos(3 * angle), np.sin(3 * angle)] for angle in angles] self.vertices = np.array(self.vertices) self.params = {'n_steps' : 3, 'vertices' : self.vertices, 'temperature' : 0.001, 'temp_cool' : 0.99, 'rand_state' : fake_random.State([])} def test_get_cycle(self): annealer = tsp_draw.base.Annealer(**self.params) true_cycle = np.concatenate([self.vertices, [self.vertices[0]]], axis = 0) test_cycle = annealer.get_cycle() np.testing.assert_equal(true_cycle, test_cycle) def test_get_energy(self): annealer = tsp_draw.base.Annealer(**self.params) cycle = annealer.get_cycle() diffs = cycle[1:] - cycle[:-1] true_energy = np.linalg.norm(diffs, axis = 1).sum() test_energy = annealer.get_energy() self.assertEqual(true_energy, test_energy) def test_update_state(self): annealer = tsp_draw.base.Annealer(**self.params) annealer._update_state() self.assertEqual(annealer.temperature, self.params['temperature'] * self.params['temp_cool']) self.assertEqual(annealer.steps_processed, 1) def test_run_proposal_trial(self): uniform_results = np.linspace(0, 1.0, 10) uniform_stack = list(np.flip(uniform_results)) params = self.params.copy() params['rand_state'] = fake_random.State(uniform_stack) annealer = tsp_draw.base.Annealer(**params) energy_diff = 0.5 * annealer.temperature critical_val = np.exp(-energy_diff / annealer.temperature) test_trials = [annealer._run_proposal_trial(energy_diff) for _ in uniform_results] true_trials = [prob < critical_val for prob in uniform_results] self.assertEqual(test_trials, true_trials) def test_do_warm_restart(self): annealer = tsp_draw.base.Annealer(**self.params) annealer.steps_processed = 5 annealer.do_warm_restart() self.assertEqual(annealer.steps_processed, 0) if __name__ == '__main__': unittest.main()
[ "mwmcgonagle@gmail.com" ]
mwmcgonagle@gmail.com
ad01571f6a61ceda4ceecc3b05110483c0b3596f
99172548909dfea0095f27d590277b87b61d8bac
/Programmers/Level1/직사각형 별찍기.py
4828430b91f5d67edafdd69237b146947bc2038e
[]
no_license
rheehot/Algorithm-53
3612e51aee15fa4bc6fecd6d2c18fa45f5572117
2651f95ca25645943a8c3fe40662092b4925f06b
refs/heads/master
2023-06-24T05:35:46.465422
2021-07-22T14:45:15
2021-07-22T14:45:15
null
0
0
null
null
null
null
UTF-8
Python
false
false
85
py
a, b = map(int, input().strip().split(' ')) rect = ("*" * a + '\n') * b print(rect)
[ "59171154+HelloJihyoung@users.noreply.github.com" ]
59171154+HelloJihyoung@users.noreply.github.com
184e8e9933bf4850ac425bc2697124c4c5776379
03c9cd5bd96874d6117fb17c37ac4d7450c15933
/Opencv-Python/wechat/autojump.py
540e6d96cb2fd16283ba2e25403877731481716d
[]
no_license
atiger808/opencv-tutorial
603de35e97679d6beae104298ae355edfdd9036a
2ea9bb3818284fb75f85697e36fde37b6479d1c6
refs/heads/master
2020-05-29T23:16:30.462022
2019-11-05T10:08:20
2019-11-05T10:08:20
189,425,004
0
0
null
null
null
null
UTF-8
Python
false
false
3,751
py
# _*_ coding: utf-8 _*_ # @Time : 2019/4/3 16:45 # @Author : Ole211 # @Site : # @File : autojump.py # @Software : PyCharm import cv2 import numpy as np import os import time import subprocess import math # os.chdir('d:\\img\\') press_coefficient = 1.35 def get_center_coord(img): region_lower = int(img.shape[0]*0.3) region_upper = int(img.shape[0]*0.7) region = img[region_lower:region_upper] hsv_img = cv2.cvtColor(region, cv2.COLOR_BGR2HSV) color_lower = np.array([105, 25, 45]) color_upper = np.array([135, 125, 130]) color_mask = cv2.inRange(hsv_img, color_lower, color_upper) _, contours, hierarchy = cv2.findContours(color_mask, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE) if len(contours)>0: max_contour = max(contours, key=cv2.contourArea) rect = cv2.boundingRect(max_contour) x, y, w, h = rect cv2.rectangle(region, (x, y), (x+w, y+h), (0, 255, 0), 3) center_coord = (x+int(w/2), y+h-20) cv2.circle(region, center_coord, 8, (0, 0, 255), -1) cv2.drawContours(region, max_contour, -1, (0, 0, 255), 3) # region = cv2.resize(region, (400, 800)) # cv2.imshow('color_mask', color_mask) # cv2.imshow('region', region) # cv2.waitKey() # cv2.destroyAllWindows() return hsv_img, color_mask, center_coord def get_box_center(img): region_lower = int(img.shape[0] * 0.3) region_upper = int(img.shape[0] * 0.7) region = img[region_lower:region_upper] gray_img = cv2.cvtColor(region, cv2.COLOR_BGR2GRAY) # cv2.imshow('gray', gray_img) canny_img = cv2.Canny(gray_img, 75, 150) y_top = np.nonzero([max(row) for row in canny_img[:400]])[0][0] x_top = int(np.mean(np.nonzero(canny_img[y_top]))) y_bottom = y_top + 200 # for row in range(y_bottom, 768): # if canny_img[row, x_top] != 0: # break box_center_coord = (x_top, (y_top + y_bottom)//2) cv2.circle(region, box_center_coord, 8, (0, 0, 255), -1) return canny_img, region, box_center_coord def pullScreenshot(): os.system('adb shell screencap -p /sdcard/autojump.png') os.system('adb pull /sdcard/autojump.png .') def jump(distance): press_time = distance * 1.35 press_time = int(press_time) cmd = 'adb shell input swipe 320 410 320 410 ' + str(press_time) print(cmd) os.system(cmd) def beginJump(): while True: pullScreenshot() time.sleep(2) img = cv2.imread('autojump.png') color_mask, hsv_img, center_coord = get_center_coord(img) canny_img, region, box_center_coord = get_box_center(img) distance = math.sqrt((box_center_coord[0] - center_coord[0]) ** 2 + (box_center_coord[1] - center_coord[1]) ** 2) w, h = region.shape[:2] text = 'press time: %.3f ms' %(max(1.35*distance, 200)) cv2.putText(region, text, (0, 50), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 1, cv2.LINE_AA) text0 = 'distance: %.3f pixels' % (distance) cv2.putText(region, text0, (0, 100), cv2.FONT_HERSHEY_SIMPLEX, 1, (0, 0, 255), 1, cv2.LINE_AA) cv2.line(region, center_coord, box_center_coord, (0, 0, 255), 3) print('棋子坐标:', center_coord) print('盒子坐标:', box_center_coord) print('距离:', distance) cv2.imwrite('region.png', region) # cv2.imshow('color_mask', color_mask) # cv2.imshow('hsv_img', hsv_img) # cv2.imshow('canny_img', canny_img) # cv2.imshow('region', region) # cv2.waitKey() # cv2.destroyAllWindows() jump(distance) time.sleep(0.2) if __name__ == '__main__': beginJump() # pullScreenshot() # if __name__ == '__main__': # get_center_coord()
[ "atiger0614@163.com" ]
atiger0614@163.com
dae76b5a56cfbe512236e47e5b204fddff746a73
4e382ae46cf997ea2dbdfcfa463a57d3e0e9ad97
/sols/gray_code.py
490b72b7d1576b6786b190e757dfced57e83460c
[]
no_license
hayeonk/leetcode
5136824838eb17ed2e4b7004301ba5bb1037082f
6485f8f9b5aa198e96fbb800b058d9283a28e4e2
refs/heads/master
2020-04-28T03:37:16.800519
2019-06-01T14:34:45
2019-06-01T14:34:45
174,943,756
0
1
null
null
null
null
UTF-8
Python
false
false
381
py
class Solution(object): def grayCode(self, n): def getCode(n): if n == 0: return ["0"] rest = getCode(n-1) reverse = reversed(rest) ret = [x + "0" for x in rest] + [x + "1" for x in reverse] return ret ret = getCode(n) ret = [int(x, 2) for x in ret] return ret
[ "31617695+hayeonk@users.noreply.github.com" ]
31617695+hayeonk@users.noreply.github.com
7fb1bbcd1838101abf13096c7d71cc1156bf7060
e3f2a0acc79f1891b93553ee6a95396edeb6fd60
/setup.py
c9c0390cc3d9d040b7b7fc777d3544fa322b0332
[ "Apache-2.0" ]
permissive
imaginal/openprocurement.search_plugins
5bd23b7e17365abba9f7f33da7a5c3263c440453
a32a5e1b54c9b02fe24fae93e2e78632f77be82a
refs/heads/master
2020-04-11T23:30:41.704868
2018-12-17T18:31:07
2018-12-17T18:31:07
162,170,079
0
0
null
null
null
null
UTF-8
Python
false
false
906
py
from setuptools import setup, find_packages setup( name='openprocurement.search_plugins', version='0.1', # NOQA description="Plugin for OpenProcurement Search", long_description=open("README.md").read(), # Get more strings from # http://pypi.python.org/pypi?:action=list_classifiers classifiers=[ "License :: OSI Approved :: Apache Software License", "Programming Language :: Python", ], keywords='prozorro search plugin', author='Volodymyr Flonts', author_email='flyonts@gmail.com', license='Apache License 2.0', url='https://github.com/imaginal/openprocurement.search_plugins', namespace_packages=['openprocurement'], packages=find_packages(), package_data={'': ['*.md', '*.txt']}, include_package_data=True, zip_safe=False, install_requires=[ 'openprocurement.search' ], entry_points={ } )
[ "flyonts@gmail.com" ]
flyonts@gmail.com
cac6679df85dec7cd9809210f80e1ddda0f67e88
1b2aeb34086ffd631e630008ffc73b6f1c458e8d
/kfold.py
4ed9cc4e0681f14208c53012d2e16279b953e4d0
[]
no_license
thunderboom/tecent_ad
bdde7af24aacce2cbe86d02e9defefa9378e3a3b
5c051ddd63848c767ecbc118841a953178b4111a
refs/heads/master
2022-09-08T02:31:49.090314
2020-05-26T08:41:48
2020-05-26T08:41:48
265,536,465
0
0
null
null
null
null
UTF-8
Python
false
false
1,198
py
"""根据切分形式,进行训练并验证""" from torch.utils.data import DataLoader from utils import AdvData from sklearn.model_selection import train_test_split from train_eval import model_train, model_evaluate def cross_validation(config, model, train_data, test_data): if config.pattern == 'cross_validation': train_data, val_data = train_test_split(train_data, test_size=config.val_size, random_state=config.seed) #分训练集和验证集 train_data, val_data, test_data = \ AdvData(train_data, config), AdvData(val_data, config), AdvData(test_data, config) train_loader = DataLoader(train_data, shuffle=True, batch_size=config.batch_size) val_loader = DataLoader(val_data, shuffle=False, batch_size=config.batch_size) test_loader = DataLoader(test_data, shuffle=False, batch_size=config.batch_size) model_trained = model_train(config, model, train_loader, val_loader) #训练模型 predict_label = None if config.test == True: predict_label = model_evaluate(config, model_trained, test_loader, test=True) #对测试集进行输出 return model_trained, predict_label
[ "470810634@qq.com" ]
470810634@qq.com
19ebe52e3014d768bb3a0b4bde461281fb6ffc50
c04b004558beee584817c26aaf4809cb366f7fbb
/Course/Section-1/day12/2.作业.py
c45cc33e1a3d05734ee700af48c634dfd2609f76
[]
no_license
Wuzhibin05/python-course
4f3d3e39990a613d222b1e7bcb3bb1453b0541a0
50daa1405b873e9c6727598e6c752c115fe7843a
refs/heads/master
2023-07-14T07:17:19.054495
2021-08-26T01:22:13
2021-08-26T01:22:13
353,526,567
0
0
null
null
null
null
UTF-8
Python
false
false
2,575
py
# 1.编写装饰器,为多个函数加上认证的功能(用户的账号密码来源于文件), # 要求登录成功一次,后续的函数都无需再输入用户名和密码 # FLAG = False # def login(func): # def inner(*args,**kwargs): # global FLAG # '''登录程序''' # if FLAG: # ret = func(*args, **kwargs) # func是被装饰的函数 # return ret # else: # username = input('username : ') # password = input('password : ') # if username == 'boss_gold' and password == '22222': # FLAG = True # ret = func(*args,**kwargs) #func是被装饰的函数 # return ret # else: # print('登录失败') # return inner # # @login # def shoplist_add(): # print('增加一件物品') # # @login # def shoplist_del(): # print('删除一件物品') # # shoplist_add() # shoplist_del() # 2.编写装饰器,为多个函数加上记录调用功能,要求每次调用函数都将被调用的函数名称写入文件 # def log(func): # def inner(*args,**kwargs): # with open('log','a',encoding='utf-8') as f: # f.write(func.__name__+'\n') # ret = func(*args,**kwargs) # return ret # return inner # # @log # def shoplist_add(): # print('增加一件物品') # # @log # def shoplist_del(): # print('删除一件物品') # shoplist_add() # shoplist_del() # shoplist_del() # shoplist_del() # shoplist_del() # shoplist_del() # 进阶作业(选做): # 1.编写下载网页内容的函数,要求功能是:用户传入一个url,函数返回下载页面的结果 # 2.为题目1编写装饰器,实现缓存网页内容的功能: # 具体:实现下载的页面存放于文件中,如果文件内有值(文件大小不为0),就优先从文件中读取网页内容,否则,就去下载,然后存到文件中 import os from urllib.request import urlopen def cache(func): def inner(*args,**kwargs): if os.path.getsize('web_cache'): with open('web_cache','rb') as f: return f.read() ret = func(*args,**kwargs) #get() with open('web_cache','wb') as f: f.write(b'*********'+ret) return ret return inner @cache def get(url): code = urlopen(url).read() return code # {'网址':"文件名"} ret = get('http://www.baidu.com') print(ret) ret = get('http://www.baidu.com') print(ret) ret = get('http://www.baidu.com') print(ret)
[ "wuzb@szkingdom.com" ]
wuzb@szkingdom.com
f8fd22a6064e8366f38550ad3d3bfbd6d473f4c1
0f187b981948b7bb1739c22b7b1d38f6c0fd4bb5
/venv/bin/django-admin.py
a2eb0df4316c4287a519a1d266d93dd3836ecd12
[]
no_license
marcosdefina/django-studies
439680e619e50e872497c44e99d8d29e441a5741
1156c891954a13037efff776779ca86c0924c8de
refs/heads/master
2020-05-18T21:24:08.371284
2019-05-02T22:29:54
2019-05-02T22:29:54
184,661,514
0
0
null
null
null
null
UTF-8
Python
false
false
210
py
#!/media/marcosdefina/WindowsBackup/Users/marki/Documents/Estonia/Python/rp-portifolio/venv/bin/python3 from django.core import management if __name__ == "__main__": management.execute_from_command_line()
[ "marcosdefina@gmail.com" ]
marcosdefina@gmail.com
192d3de4d4523842cca9e342f1b2e292e06c8306
0a921dfd9e627153552b62170eba0f92a817ce5d
/tests/auger/test_configs.py
b0d656ecbe56dd70adff2abca21760be807d92de
[ "Apache-2.0" ]
permissive
mha23/a2ml
b7507729f3c54b71af8a1e4595c33563062d515e
8fbd1be7a81be5f41564debeecb61379097f6fb8
refs/heads/master
2021-01-02T21:25:59.063460
2019-12-11T10:04:48
2019-12-11T10:04:48
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,349
py
from .mock_rest_api import interceptor from auger.api.cloud.experiment import AugerExperimentApi EXPERIMENT = { 'data': { 'name': 'iris-1.csv-experiment', 'project_file_id': 1256, } } PROJECT_FILE = { 'data': { 'name': 'iris-1.csv', 'id': 1256, 'statistics': { 'columns_count': 2, 'count': 150, 'stat_data': [{ 'datatype': 'categorical', 'column_name': 'species', 'unique_values': 3 },{ 'datatype': 'integer', 'column_name': 'sepal_length' },{ 'datatype': 'integer', 'column_name': 'sepal_width' },{ 'datatype': 'integer', 'column_name': 'petal_length' },{ 'datatype': 'integer', 'column_name': 'petal_width' }] }, } } class TestConfigs(object): def test_experiment_settings(self, project, ctx, monkeypatch): config = ctx.get_config('auger') config.experiment.cross_validation_folds = 55 config.experiment.max_total_time = 606 config.experiment.max_eval_time = 55 config.experiment.max_n_trials = 101 config.experiment.use_ensemble = False PAYLOAD = { 'get_experiment': EXPERIMENT, 'get_project_file': PROJECT_FILE } interceptor(PAYLOAD, monkeypatch) config, model_type = AugerExperimentApi( ctx, 'project-api', 'iris-1.csv-experiment', '1234').\ get_experiment_settings() assert config['evaluation_options']['crossValidationFolds'] == 55 assert config['evaluation_options']['max_total_time_mins'] == 606 assert config['evaluation_options']['max_eval_time_mins'] == 55 assert config['evaluation_options']['max_n_trials'] == 101 assert config['evaluation_options']['use_ensemble'] == False # dataset assert config['evaluation_options']['targetFeature'] == 'species' assert config['evaluation_options']['featureColumns'] == \ ['sepal_length', 'sepal_width', 'petal_length', 'petal_width'] assert config['evaluation_options']['categoricalFeatures'] == \ ['species'] assert config['evaluation_options']['timeSeriesFeatures'] == [] assert config['evaluation_options']['binaryClassification'] == False assert config['evaluation_options']['labelEncodingFeatures'] == [] assert config['evaluation_options']['classification'] == True assert config['evaluation_options']['scoring'] == 'f1_macro' def test_exclude_setting(self, project, ctx, monkeypatch): config = ctx.get_config('config') config.exclude = ['sepal_length'] PAYLOAD = { 'get_experiment': EXPERIMENT, 'get_project_file': PROJECT_FILE } interceptor(PAYLOAD, monkeypatch) config, model_type = AugerExperimentApi( ctx, 'project-api', 'iris-1.csv-experiment', '1234').\ get_experiment_settings() assert config['evaluation_options']['targetFeature'] == 'species' assert config['evaluation_options']['featureColumns'] == \ ['sepal_width', 'petal_length', 'petal_width'] assert config['evaluation_options']['categoricalFeatures'] == \ ['species'] def test_model_type_setting(self, project, ctx, monkeypatch): config = ctx.get_config('config') config.model_type = 'regression' config = ctx.get_config('auger') config.experiment.metric = None PAYLOAD = { 'get_experiment': EXPERIMENT, 'get_project_file': PROJECT_FILE } interceptor(PAYLOAD, monkeypatch) config, model_type = AugerExperimentApi( ctx, 'project-api', 'iris-1.csv-experiment', '1234').\ get_experiment_settings() assert config['evaluation_options']['timeSeriesFeatures'] == [] assert config['evaluation_options']['binaryClassification'] == False assert config['evaluation_options']['labelEncodingFeatures'] == [] assert config['evaluation_options']['classification'] == False assert config['evaluation_options']['scoring'] == 'r2'
[ "vlad@dplrn.com" ]
vlad@dplrn.com
6ffd580688dd7af218faa5a26e83fabfce825f2a
300993f8ab336b19799ad10c7520514db6d56d2f
/aafield/aafieldapp/models.py
713013d3b46a057dd3b6beedefd197efd2d46880
[]
no_license
pratheebapalanisami/ParkReservationSystem
23ef1a42690f7697e8b804ab2ac1714128adcbd0
5bd436c1e3688b8bdcd4986135604250240c5f6e
refs/heads/master
2022-11-30T00:34:47.598881
2020-08-08T20:48:09
2020-08-08T20:48:09
286,116,956
0
0
null
null
null
null
UTF-8
Python
false
false
3,437
py
from django.contrib.auth.models import AbstractUser from django.db import models from django.utils import timezone from django.contrib.auth import get_user_model from django.urls import reverse from accounts.models import Profile from aafield import settings class Parks(models.Model): Park_Name = models.CharField(max_length=50,blank=False, null=False, default=' ') Park_Address = models.CharField(max_length=250, blank=True, null=True, default=' ') County = models.CharField(max_length=50, blank=True, null=True, default=' ') Park_Image = models.ImageField(default='Images/park2.jpg') def __str__(self): return self.Park_Name # Create your models here. class Park_Properties(models.Model): Park_Name = models.ForeignKey(Parks, on_delete=models.CASCADE) Property_Name = models.CharField(max_length=50, default=' ', null=True, blank=True) Property_Description = models.CharField(max_length=50, default=' ', null=True, blank=True) Property_Guest_Capacity = models.IntegerField(blank=True, null=True) Location_Choices=[('Indoor' ,'Indoor'),('Outdoor','Outdoor')] Property_Location= models.CharField(max_length=50,default='Indoor',choices=Location_Choices) Slot= models.CharField(max_length=50, default='2 hours', null=True, blank=True) Price = models.IntegerField(blank=True, null=True) Property_Image=models.ImageField(default='Images/pool.jpg') def __str__(self): return self.Property_Name class Reservation(models.Model): Park_Name = models.ForeignKey(Parks, on_delete=models.CASCADE) Property_Name = models.ForeignKey(Park_Properties, on_delete=models.CASCADE) Customer_Name = models.ForeignKey('accounts.customuser', on_delete=models.CASCADE) Event_Date = models.DateField(blank=True, null=True) Slot = models.CharField(max_length=50, default=' ', null=True, blank=True) Team_Size = models.CharField(max_length=50, default=' ', null=True, blank=True) Status = models.CharField(max_length=50, default=' ', null=True, blank=True) def __str__(self): return self.Property_Name.Property_Name class Property_Status(models.Model): Park_Name = models.ForeignKey(Parks, on_delete=models.CASCADE) Property_Name = models.ForeignKey(Park_Properties, on_delete=models.CASCADE) Report_TimeDate = models.DateTimeField(blank=True, null=True) Property_Status_Description = models.CharField(max_length=50, blank=True, null=True) Expenses = models.CharField(max_length=50, default=' ', null=True, blank=True) Maintenance_ID = models.ForeignKey('accounts.customuser', on_delete=models.CASCADE) Reservation_ID=models.ForeignKey(Reservation,on_delete=models.CASCADE,related_name='reservations') def __str__(self): return self.Park_Name.Park_Name class Transaction(models.Model): Park_Name = models.ForeignKey(Parks, on_delete=models.CASCADE) Property_Name = models.ForeignKey(Park_Properties, on_delete=models.CASCADE) Reservation_ID = models.ForeignKey(Reservation, on_delete=models.CASCADE) Trans_Amount = models.CharField(max_length=50, default=' ', null=True, blank=True) Trans_Time_Date = models.DateTimeField(blank=True, null=True) Trans_Type = models.CharField(max_length=50, default=' ', null=True, blank=True) Transaction_Token = models.CharField(max_length=50, default=' ', null=True, blank=True) def __str__(self): return self.Trans_Amount
[ "54645809+pratheebapalanisami@users.noreply.github.com" ]
54645809+pratheebapalanisami@users.noreply.github.com
5b87225c7d50e4de10c2e589bdd51d172792eef9
e9b3842b3d2946bb6ddc07fcb8410732ca76f613
/mdl.py
67f802dd819229e1e48f72eebcab1278ee6efc6d
[]
no_license
aditihaiman/final-graphics
2187c48bf3516467996eb1264019b0d2ea393055
45ee0b9c2733677a45d787a56f9a0333de2ad0a7
refs/heads/master
2022-11-09T00:31:56.646702
2020-06-11T22:04:50
2020-06-11T22:04:50
265,695,784
0
1
null
null
null
null
UTF-8
Python
false
false
15,997
py
from ply import lex, yacc tokens = ( "STRING", "ID", "XYZ", "DOUBLE", "INT", "COMMENT", "LIGHT", "CONSTANTS", "SAVE_COORDS", "CAMERA", "AMBIENT", "CYLINDER", "PYRAMID", "PRISM", "TORUS", "SPHERE", "ELLIPSOID", "TUBE", "CONE", "BOX", "LINE", "MESH", "TEXTURE", "SET", "MOVE", "SCALE", "ROTATE", "BASENAME", "SAVE_KNOBS", "TWEEN", "FRAMES", "VARY", "PUSH", "POP", "SAVE", "GENERATE_RAYFILES", "SHADING", "SHADING_TYPE", "SET_KNOBS", "FOCAL", "DISPLAY", "SCREEN", "WEB", "CO" ) reserved = { "x" : "XYZ", "y" : "XYZ", "z" : "XYZ", "screen" : "SCREEN", "light" : "LIGHT", "constants" : "CONSTANTS", "save_coord_system" : "SAVE_COORDS", "camera" : "CAMERA", "ambient" : "AMBIENT", "cylinder" : "CYLINDER", "pyramid" : "PYRAMID", "torus" : "TORUS", "sphere" : "SPHERE", "ellipsoid" : "ELLIPSOID", "cone" : "CONE", "tube" : "TUBE", "prism" : "PRISM", "box" : "BOX", "line" : "LINE", "mesh" : "MESH", "texture" : "TEXTURE", "set" : "SET", "move" : "MOVE", "scale" : "SCALE", "rotate" : "ROTATE", "basename" : "BASENAME", "save_knobs" : "SAVE_KNOBS", "tween" : "TWEEN", "frames" : "FRAMES", "vary" : "VARY", "push" : "PUSH", "pop" : "POP", "save" : "SAVE", "generate_rayfiles" : "GENERATE_RAYFILES", "shading" : "SHADING", "phong" : "SHADING_TYPE", "flat" : "SHADING_TYPE", "gouraud" : "SHADING_TYPE", "raytrace" : "SHADING_TYPE", "wireframe" : "SHADING_TYPE", "setknobs" : "SET_KNOBS", "focal" : "FOCAL", "display" : "DISPLAY", "web" : "WEB" } t_ignore = " \t" def t_ID(t): r'[a-zA-Z_][a-zA-Z_0-9]*' if t.value in reserved: t.type = reserved.get(t.value) return t def t_STRING(t): r'\.[a-zA-Z_0-9]*[a-zA-Z_][a-zA-Z_0-9]*' return t def t_DOUBLE(t): r"""\-?\d+\.\d*|\-?\.\d+ | \-?\d+""" t.value = float(t.value) return t def t_COMMENT(t): r"//.*" return t def t_CO(t): r":" return t def t_error(t): print("TOKEN ERROR: " + str(t)) lex.lex() #---------------------------------------------------------- commands = [] symbols = {} def p_input(p): """input : | command input""" pass def p_command_comment(p): 'command : COMMENT' pass def p_SYMBOL(p): """SYMBOL : XYZ | ID""" p[0] = p[1] def p_TEXT(p): """TEXT : SYMBOL | STRING""" p[0] = p[1] def p_NUMBER(p): """NUMBER : DOUBLE""" p[0] = p[1] def p_command_stack(p): """command : POP | PUSH""" commands.append({'op' : p[1], 'args' : None}) def p_command_screen(p): """command : SCREEN NUMBER NUMBER | SCREEN""" if len(p) == 2: commands.append({'op' : p[1], 'width' : 500, 'height': 500}) else: commands.append({'op' : p[1], 'width' : p[2], 'height': p[3]}) def p_command_save(p): """command : SAVE TEXT TEXT""" commands.append({'op' : p[1], 'args' : [p[2]]}) def p_command_show(p): """command : DISPLAY""" commands.append({'op' : p[1], 'args' : None}) def p_command_cylinder(p): """command : CYLINDER NUMBER NUMBER NUMBER NUMBER NUMBER | CYLINDER SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER | CYLINDER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | CYLINDER SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 8 and isinstance(p[7], str): cmd['cs'] = p[7] if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] cmd['args'] = p[arg_start:arg_start+5] commands.append(cmd) def p_command_tube(p): """command : TUBE NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | TUBE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | TUBE NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | TUBE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] if len(p) == 10 and isinstance(p[9], str): cmd['cs'] = p[9] cmd['args'] = p[arg_start:arg_start+6] commands.append(cmd) def p_command_cone(p): """command : CONE NUMBER NUMBER NUMBER NUMBER NUMBER | CONE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER | CONE NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | CONE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 8 and isinstance(p[7], str): cmd['cs'] = p[7] if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] cmd['args'] = p[arg_start:arg_start+5] commands.append(cmd) def p_command_prism(p): """command : PRISM NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | PRISM SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | PRISM NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | PRISM SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] if len(p) == 10 and isinstance(p[9], str): cmd['cs'] = p[9] cmd['args'] = p[arg_start:arg_start+6] commands.append(cmd) def p_command_pyramid(p): """command : PYRAMID NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | PYRAMID SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | PYRAMID NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | PYRAMID SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 11 and isinstance(p[10], str): cmd['cs'] = p[10] if len(p) == 12 and isinstance(p[11], str): cmd['cs'] = p[11] cmd['args'] = p[arg_start:arg_start+8] commands.append(cmd) def p_command_ellipsoid(p): """command : ELLIPSOID NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | ELLIPSOID SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | ELLIPSOID NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | ELLIPSOID SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" #print("P", p[7]) cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] if len(p) == 10 and isinstance(p[9], str): cmd['cs'] = p[9] cmd['args'] = p[arg_start:arg_start+6] commands.append(cmd) def p_command_sphere(p): """command : SPHERE NUMBER NUMBER NUMBER NUMBER | SPHERE SYMBOL NUMBER NUMBER NUMBER NUMBER | SPHERE NUMBER NUMBER NUMBER NUMBER SYMBOL | SPHERE SYMBOL NUMBER NUMBER NUMBER NUMBER SYMBOL""" #print("P", p[7]) cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 7 and isinstance(p[6], str): cmd['cs'] = p[6] if len(p) == 8 and isinstance(p[7], str): cmd['cs'] = p[7] cmd['args'] = p[arg_start:arg_start+4] commands.append(cmd) def p_command_torus(p): """command : TORUS NUMBER NUMBER NUMBER NUMBER NUMBER | TORUS NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | TORUS SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER | TORUS SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 8 and isinstance(p[7], str): cmd['cs'] = p[7] if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] cmd['args'] = p[arg_start:arg_start+5] commands.append(cmd) def p_command_box(p): """command : BOX NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | BOX NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | BOX SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | BOX SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 if len(p) == 9 and isinstance(p[8], str): cmd['cs'] = p[8] if len(p) == 10 and isinstance(p[9], str): cmd['cs'] = p[9] cmd['args'] = p[arg_start:arg_start+6] commands.append(cmd) def p_command_line(p): """command : LINE NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | LINE NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | LINE NUMBER NUMBER NUMBER SYMBOL NUMBER NUMBER NUMBER | LINE NUMBER NUMBER NUMBER SYMBOL NUMBER NUMBER NUMBER SYMBOL | LINE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | LINE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER SYMBOL | LINE SYMBOL NUMBER NUMBER NUMBER SYMBOL NUMBER NUMBER NUMBER | LINE SYMBOL NUMBER NUMBER NUMBER SYMBOL NUMBER NUMBER NUMBER SYMBOL""" cmd = {'op' : p[1], 'constants' : None, 'cs0' : None, 'cs1' : None, 'args':[]} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start = 3 cmd['args'] = p[arg_start:arg_start+3] arg_start = arg_start+3 if isinstance(p[arg_start], str): cmd['cs0'] = p[arg_start] arg_start+= 1 cmd['args']+= p[arg_start:arg_start+3] if len(p) == 9 and isinstance(p[8], str): cmd['cs1'] = p[8] if len(p) == 10 and isinstance(p[9], str): cmd['cs1'] = p[9] if len(p) == 11 and isinstance(p[10], str): cmd['cs1'] = p[10] commands.append(cmd) def p_command_move(p): """command : MOVE NUMBER NUMBER NUMBER SYMBOL | MOVE NUMBER NUMBER NUMBER""" cmd = {'op' : p[1], 'args' : p[2:5], 'knob' : None} if len(p) == 6: cmd['knob'] = p[5] symbols[p[5]] = ['knob', 0] commands.append(cmd) def p_command_scale(p): """command : SCALE NUMBER NUMBER NUMBER SYMBOL | SCALE NUMBER NUMBER NUMBER""" cmd = {'op' : p[1], 'args' : p[2:5], 'knob' : None} if len(p) == 6: cmd['knob'] = p[5] symbols[p[5]] = ['knob', 0] commands.append(cmd) def p_command_rotate(p): """command : ROTATE XYZ NUMBER SYMBOL | ROTATE XYZ NUMBER""" cmd = {'op' : p[1], 'args' : p[2:4], 'knob' : None} if len(p) == 5: cmd['knob'] = p[4] symbols[p[4]] = ['knob', 0] commands.append(cmd) def p_command_frames(p): """command : FRAMES NUMBER""" cmd = {'op' : p[1], 'args' : [p[2]]} commands.append(cmd) def p_command_basename(p): """command : BASENAME TEXT""" cmd = {'op' : p[1], 'args' : [p[2]]} commands.append(cmd) def p_command_vary(p): """command : VARY SYMBOL NUMBER NUMBER NUMBER NUMBER""" cmd = {'op' : p[1], 'args' : p[3:], 'knob' : p[2]} symbols[p[2]] = ['knob', 0] commands.append(cmd) def p_command_knobs(p): """command : SET SYMBOL NUMBER | SET_KNOBS NUMBER""" cmd = {'op' : p[1], 'args' : [], 'knob' : None} if p[1] == 'SET': cmd['knob'] = p[2] cmd['args'].append(p[3]) symbols[p[2]] = p[3] else: cmd['args'].append(p[2]) commands.append(cmd) def p_command_ambient(p): "command : AMBIENT NUMBER NUMBER NUMBER" symbols['ambient'] = ['ambient'] + p[2:] cmd = {'op':p[1], 'args':p[2:]} commands.append(cmd) def p_command_constants(p): """command : CONSTANTS SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER | CONSTANTS SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER""" symbols[p[2]] = ['constants', {'red' : p[3:6], 'green' : p[6:9], 'blue' : p[9:]}] cmd = {'op':p[1], 'args' : None, 'constants' : p[2] } commands.append(cmd) def p_command_light(p): "command : LIGHT SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER" symbols[p[2]] = ['light', {'location' : p[3:6], 'color' : p[6:]}] cmd = {'op':p[1], 'args' : None, 'light' : p[2] } commands.append(cmd) def p_command_shading(p): "command : SHADING SHADING_TYPE" symbols['shading'] = ['shade_type', p[2]] cmd = {'op':p[1], 'args' : None, 'shade_type' : p[2] } commands.append(cmd) def p_command_camera(p): "command : CAMERA NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER" symbols['camera'] = ['camera', {'eye': p[2:4], 'aim': p[4:]} ] commands.append({'op':p[1], 'args':None}) def p_command_generate_rayfiles(p): "command : GENERATE_RAYFILES" commands.append({'op':p[1], 'args':None}) def p_command_mesh(p): """command : MESH CO TEXT | MESH SYMBOL CO TEXT | MESH CO TEXT SYMBOL | MESH SYMBOL CO TEXT SYMBOL""" cmd = {'op':p[1], 'args' : [], 'cs':None, 'constants':None} arg_start = 2 if isinstance(p[2], str): cmd['constants'] = p[2] arg_start+= 1 cmd['args'].append(p[arg_start]) if len(p) == 4 and isinstance(p[3], str): cmd['cs'] = p[3] if len(p) == 5 and isinstance(p[4], str): cmd['cs'] = p[4] commands.append(cmd) def p_save_knobs(p): "command : SAVE_KNOBS SYMBOL" cmd = {'op':p[1], 'args':None, 'knob_list':p[2]} symbols[p[2]] = ['knob_list', []] commands.append(cmd) def p_save_coords(p): "command : SAVE_COORDS SYMBOL" cmd = {'op':p[1], 'args':None, 'cs':p[2]} symbols[p[2]] = ['coord_sys', []] commands.append(cmd) def p_tween(p): "command : TWEEN NUMBER NUMBER SYMBOL SYMBOL" cmd = {'op':p[1], 'args':p[2:4], 'knob_list0':p[4], 'knob_list1':p[5]} commands.append(cmd) def p_focal(p): "command : FOCAL NUMBER" commands.append({'op':p[1], 'args':[p[2]]}) def p_web(p): "command : WEB" commands.append({'op':p[1], 'args':None}) def p_texture(p): "command : TEXTURE SYMBOL NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER NUMBER" symbols[p[2]] = ['texture', p[3:]] def p_error(p): print('SYNTAX ERROR: ' + str(p)) yacc.yacc() from copy import deepcopy def parseFile(filename): """ This function returns a tuple containing a list of opcodes and a list of symbols. Every opcode is a tuple of the form (commandname, parameter, parameter, ...). Every symbol is a tuple of the form (type, name). """ global commands global symbols commands = [] symbols = {} try: f = open(filename, "r") for line in f.readlines(): line = line.strip() yacc.parse(line) f.close() result = (commands[:], deepcopy(symbols)) commands = [] symbols = {} return result except IOError: return ()
[ "ahaiman00@stuy.edu" ]
ahaiman00@stuy.edu
8beec3595266a31881890de7dda1c3111fee3c79
f0237dd96510765f1a8d4b61e5e6dca0d1142fba
/lib/tools/tools_ssh.py
f6155f4fca7f6f97a9800cfdbb16671dbf2a4638
[ "MIT" ]
permissive
galena503/SCR
d217782dd03f22f5524666f11ccd6c9e73bec223
d5b6581808b4f2fac775e7ff48b3eef548164ca1
refs/heads/master
2020-08-01T06:56:33.715559
2019-09-26T09:34:45
2019-09-26T09:34:45
210,895,454
0
0
null
null
null
null
UTF-8
Python
false
false
151
py
class Tools_ssh: def get_other_scr_stats(self): # SCRfield内の他のSCRの情報を取得する def get_other_scr_stats(self):
[ "52788168+akatuki-hi@users.noreply.github.com" ]
52788168+akatuki-hi@users.noreply.github.com
fcd7b5300758601302ad4cce5f20949aca28de24
292a5ad7aa1031a00ae8b0f4b7514fd3480ab209
/data/get_csv.py
dee2c25febe9c506dcf0ad20300fc39ff7e50463
[]
no_license
bevarb/Autofocus-for-SPRM
d3e52140cf5bd6e5e0a4c856fa6e0449042875cc
49ab4632b0097cd42501110da84ce2fa0e0b6285
refs/heads/master
2022-12-29T10:17:28.986442
2020-10-16T07:21:12
2020-10-16T07:21:12
268,509,896
1
0
null
null
null
null
UTF-8
Python
false
false
2,635
py
import pandas as pd import numpy as np import os def get_csv(B_root, A_root, save_root, CENTER=100): '''获得A、B文件夹图片的序列信息,找到对焦帧,存入csv文件''' A_List = os.listdir(A_root) A_List = sorted(A_List, key=lambda x: int(x.split("_")[0])) B_List = os.listdir(B_root) B_List = sorted(B_List, key=lambda x: int(x.split("_")[0])) Data = [["None" for _ in range(CENTER*2)] for _ in range(len(A_List))] for i in range(len(A_List)): Data[i][CENTER] = A_List[i] id = int(A_List[i].split("_")[0]) A_frame = int(A_List[i].split("_")[-1].split(".")[0]) B_path = [] for B in B_List: if int(B.split("_")[0]) == id: B_path.append(B) B_path = sorted(B_path, key=lambda x: int(x.split("_")[-1].split(".")[0])) for B in B_path: B_frame = int(B.split("_")[-1].split(".")[0]) real = A_frame - B_frame # 实际失焦帧 Data[i][CENTER - real] = B # 在对应位置上添加 nums = [] # 计算每个焦距存在的图片数量 for i in range(CENTER*2): num = 0 for j in range(len(A_List)): if Data[j][i] != "None": num += 1 elif Data[j][i] == "None": Data[j][i] = "" nums.append(num) Data.append(nums) Data = pd.DataFrame(Data, columns=[str(i) for i in range(-100, 100, 1)]) Data.to_csv(save_root + "/" + "data.csv", index=False) print(Data) def get_test_csv(root): '''获得0.1序列下的数据,所有数据都按从左到右的顺序,长短不一''' All_name = os.listdir(root + "/B") All_name = sorted(All_name, key=lambda x: int(x.split("_")[0])) L = int(All_name[-1].split("_")[0]) All_DATA = [] for i in range(L): temp = ["NONE" for i in range(100)] All_DATA.append(temp) for name in All_name: path = name x, y = int(name.split("_")[0]) - 1, int(name.split("_")[-1].split(".")[0]) - 1 print(path, x, y) All_DATA[x][y] = path # 记录时间 t = [] for i in range(100): flag = 0 for j in range(L): if All_DATA[j][i] != "NONE": flag += 1 t.append(flag) All_DATA.append(t) Data = pd.DataFrame(All_DATA, columns=[str(i) for i in range(0, 100)]) Data.to_csv(root + "/" + "newdata.csv", index=False) # get_test_csv("New-ROI-0.1/test1/new_for_train_2") B_root = "New-Total-Train/for_train/B" A_root = "New-Total-Train/for_train/A" save_root = "New-Total-Train/for_train" get_csv(B_root, A_root, save_root)
[ "49848629+bevarb@users.noreply.github.com" ]
49848629+bevarb@users.noreply.github.com
a9597573158cd06dab3973ee6e0512978f90458b
229d71da5bb9fcdc34ab9c3a4ff9f75ca7ea7a19
/bitly_app/urls.py
81c9ebb0845bfee3a27ec09d812bed36ced4f7b6
[]
no_license
Cunarefa/Convert
8fd1ba5aae46915b1cde31a682d6ddd1b83bbeef
93d366656c51affc2d17c685fcd6c93345180a49
refs/heads/master
2023-08-29T11:42:02.784981
2021-09-18T10:28:44
2021-09-18T10:28:44
407,829,331
0
0
null
2021-09-20T18:39:30
2021-09-18T10:30:09
Python
UTF-8
Python
false
false
133
py
from django.urls import path from .views import ConvertView urlpatterns = [ path('long', ConvertView.as_view(), name='long'), ]
[ "yevgen.yelik@gmail.com" ]
yevgen.yelik@gmail.com
8fd914c3d126ba141d422b9c77ea8058d4bed139
ffe606c85de9009d2c15356f82daa524c343b925
/11.28.cleanup/s2rarecats_prep.py
2f94abfa78be1fcaf7c2caf0858efeb8b810ee3c
[]
no_license
jbinkleyj/story_writer
d88ff7e3360fb8afd12445d1cb237788636b3083
dc5106a35f5fbce72f8cf0801c0ad4cbc0c9f12f
refs/heads/master
2020-07-09T15:54:02.492373
2017-12-16T07:26:59
2017-12-16T07:26:59
null
0
0
null
null
null
null
UTF-8
Python
false
false
177
py
from preprocess import * from arguments import s2s2catsrare as parseParams if __name__=="__main__": args = parseParams() DS = load_data(args) torch.save(DS,args.datafile)
[ "kedzior@uw.edu" ]
kedzior@uw.edu
e28d643bbf487ab16a8d7714dc8b7719cd5d4d27
8b677a1618a656a7544befb2a49b430b9e4bc236
/spriteSizeCheck.py
c241c5b1ccb0edd62cf845d486654e1f93ecbb6d
[]
no_license
charlottehill/snapshots
bb364e878e1ea3098f10dabbc9903f1d6aaf6e92
b4d12b3ef73c24bad04c47e66936ec1a0546f515
refs/heads/master
2016-08-04T13:16:41.341537
2014-04-14T09:11:28
2014-04-14T09:11:28
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,551
py
from __future__ import print_function from collections import Counter from kelp.kelpplugin import KelpPlugin from snapshotplugin import SnapshotPlugin import os import sys import kurt class SpriteSizeCheck(SnapshotPlugin): def __init__(self): super(SpriteSizeCheck, self) self._modulename = "Sprite Size Check" self._moduledescription = "Checks to see if any sprites become abnormally small" def checkSize(self, oct): results = [] for sprite in oct.sprites: if sprite.size <= 20: results.append("Size of " + sprite.name + " is very small") else: results.append("Size of " + sprite.name + " is normal") return results def analyze(self, path): results = {} dirname = os.path.dirname(path) for file in os.listdir(path): if file.endswith(".oct"): base = os.path.basename(file) filename = os.path.splitext(base)[0] oct = kurt.Project.load(os.path.abspath(os.path.join(path, file))) strings = (self.checkSize(oct)) results[filename] = '' for item in strings: results[filename] += item +', ' return results def size_display(results): output = {} count = 1 for filename, htmlout in results.items(): htmlout = '<h2>' + 'Project #' + str(count) + '</h2>' + '<p>' + str(htmlout) + '</p>' output[count] = ''.join(htmlout) count+=1 return output
[ "tmartinez@umail.ucsb.edu" ]
tmartinez@umail.ucsb.edu
4f05de413d1b0bc99e8d424f2680575e2250630a
e3365bc8fa7da2753c248c2b8a5c5e16aef84d9f
/indices/nnoilier.py
e97e8bf74b3364175a1104be69b64e8c77151a39
[]
no_license
psdh/WhatsintheVector
e8aabacc054a88b4cb25303548980af9a10c12a8
a24168d068d9c69dc7a0fd13f606c080ae82e2a6
refs/heads/master
2021-01-25T10:34:22.651619
2015-09-23T11:54:06
2015-09-23T11:54:06
42,749,205
2
3
null
2015-09-23T11:54:07
2015-09-18T22:06:38
Python
UTF-8
Python
false
false
449
py
ii = [('UnitAI.py', 4), ('LeakWTI3.py', 2), ('SeniNSP.py', 1), ('LyelCPG.py', 1), ('GilmCRS.py', 1), ('AinsWRR.py', 2), ('NewmJLP.py', 1), ('CoopJBT.py', 1), ('BachARE.py', 1), ('MereHHB3.py', 2), ('WilkJMC.py', 2), ('WestJIT.py', 1), ('LyttELD3.py', 1), ('SomeMMH.py', 1), ('HaliTBC.py', 1), ('AinsWRR2.py', 1), ('ClarGE3.py', 1), ('EvarJSP.py', 1), ('SadlMLP2.py', 1), ('BowrJMM2.py', 1), ('LyelCPG3.py', 1), ('BowrJMM3.py', 1), ('DibdTBR.py', 1)]
[ "varunwachaspati@gmail.com" ]
varunwachaspati@gmail.com
ab30ab18ba8f3c1c08c46beca037f2dfd35c0f29
df6cca824ab8c71b8c054b0c5accc24d0593d2e4
/accounts/views.py
2291198fd5f6cb2d867b0b90884894dc6da6837e
[]
no_license
emanan94/django-mostudio
ba4fda2fb4cc4af39e92b46cae92aa6fc61d32da
9178346c01df1253b61a74be01418f956c7c55c4
refs/heads/main
2023-01-11T09:33:35.165511
2020-11-13T17:59:59
2020-11-13T17:59:59
304,308,342
0
0
null
null
null
null
UTF-8
Python
false
false
1,819
py
from django.shortcuts import render, redirect from django.urls import reverse from .models import Profile from .forms import ProfileForm, UserForm,UserCreateForm from django.contrib.auth import authenticate, login #Create your views here. def signup(request): if request.method == 'POST': signup_form = UserCreateForm(request.POST) if signup_form.is_valid(): signup_form.save() # return redirect(reverse('login')) username = signup_form.cleaned_data['username'] password = signup_form.cleaned_data['password1'] user = authenticate(username=username,password=password) login(request,user) return redirect(reverse('accounts:profile')) else: signup_form = UserCreateForm() return render(request,'registration/signup.html',{'signup_form':signup_form}) def profile(request): profile=Profile.objects.get(user=request.user) return render(request,'profile/profile.html',{'profile':profile}) def profile_edit(request): profile=Profile.objects.get(user = request.user) if request.method == 'POST': user_form = UserForm(request.POST , instance=request.user) profile_form = ProfileForm(request.POST , request.FILES , instance=profile) if user_form.is_valid() and profile_form.is_valid(): user_form.save() my_form = profile_form.save(commit=False) my_form.user = request.user my_form.save() return redirect(reverse('profile/')) else: user_form = UserForm(instance=request.user) profile_form = ProfileForm(instance = profile) return render(request,'profile/profile_edit.html',{ 'user_form' : user_form , 'profile_form' : profile_form })
[ "youngamer_@outlook.com" ]
youngamer_@outlook.com
eeff077ecf7d9470bde1d09731515339125cc0b0
41738ae30e82facd15b3b77ace6bd1a0013f5296
/analytics/routes/login/baseroutes.py
ada3b240c27e9de4420421bdf7fc06b485e6680d
[]
no_license
sreenathp20/technicolor_backend
4c3bf5489ce4516face53699da630935a97fb21f
10dd323512ef3a8bad058a93b17e5c6efc203bab
refs/heads/master
2021-01-10T19:41:30.427869
2015-07-12T08:23:21
2015-07-12T08:23:21
38,954,648
0
0
null
null
null
null
UTF-8
Python
false
false
3,091
py
from analytics import app, login_user, login_required, logout_user, login_manager, current_user from flask import render_template, redirect, request, session, jsonify import simplejson as json, sys from analytics.lib.user.user import User from analytics.lib.user.manageuser import ManageUser from analytics.routes.login.configroutes import load_config from bson import json_util from os import listdir from os.path import isfile, join @app.route("/login") #@load_config def login(): #print "test", app.test print return render_template('pages/login.html') @login_manager.unauthorized_handler def unauthorized(): # do stuff print "unauthorized", current_user.get_id() return redirect("/login") @app.route("/") @login_required def home(): section = app.current_user.main_access print "current_user.main_access", app.current_user.main_access #return render_template('pages/'+section+'/layout_blank_page.html', section=section) return redirect('/'+section+''+'/') #return "under development" @app.route("/loginuser", methods=["POST"]) def loginuser(): data = json.loads(request.data) print 'data', request.data if "username" in data and "password" in data: params = {} params["username"] = data['username'] params["password"] = data['password'] u = User(params["username"], params["password"]) if u.is_authenticated(): res = {"success": True} login_user(u) mu = ManageUser() print "request.remote_addr", request.remote_addr mu.SendNotification(request.remote_addr) return json.dumps(res) else: res = {"success": False} return json.dumps(res) else: res = {"success": False} return json.dumps(res) @app.route("/get_users", methods=["POST"]) def get_users(): data = json.loads(request.data) mu = ManageUser() print 'data', data users = mu.GetUsers(data) #print 'users', users return json.dumps(users) @app.route("/get_status", methods=["POST"]) def get_status(): mu = ManageUser() status = mu.GetDbStatus() return json.dumps(status) @app.route("/get_files", methods=["POST"]) def get_files(): data = json.loads(request.data) mypath = data['path'] #mypath = '/home/sreenath/Documents/projects/backend' onlyfiles = [ f for f in listdir(mypath) if isfile(join(mypath,f)) ] print 'onlyfiles', onlyfiles return json.dumps(onlyfiles) @app.route("/logout", methods=["GET", "POST"]) @login_required def logout(): logout_user() return redirect("/login") @login_manager.user_loader def load_user(userid): #print "userid", userid try: mu = ManageUser() cred = mu.GetCredential(str(userid)) #print "cred", cred u = User(cred["_id"], cred["password"]) u.role = cred["role"] u.access = cred["access"] u.main_access = cred["main_access"] except: print "load_user error:", sys.exc_info()[0] u = "error check in log" return u
[ "sreenathp20@gmail.com" ]
sreenathp20@gmail.com
e1a6fea9523512dd08abfb5291de92885def9a2a
eff81639c9fb318697b475f6d076021b9187b3eb
/bruna_vasconcelos_E11.py
29a034f3bb52e70f9ddf7e5c405382f68a8137f2
[]
no_license
bruna/lab-de-programacao
85813a420f190df7de69b2e78283b9b6a5f59e62
0324827780fe4b4a2a91bff47a19a32feeae60c1
refs/heads/master
2021-06-01T05:49:51.710461
2016-06-20T02:07:12
2016-06-20T02:07:12
45,427,922
2
0
null
null
null
null
UTF-8
Python
false
false
5,005
py
import sys from math import floor class Node: def __init__(self, valor): self.__valor=valor self.__prox=None def getValor(self): return self.__valor def setValor(self,valor): self.__valor=valor def getProx(self): return self.__prox def setProx(self,prox): self.__prox=prox class Lista: def __init__(self): self._inicio=None self._fim=None def getInicio(self): return self._inicio.getValor() def setInicio(self,inicio): self._inicio=inicio def getFim(self): return self._fim.getValor() def setFim(self, fim): self._fim=fim def add(self, element): ad=Node(element) if self._inicio==None: self._inicio=ad self._fim=ad else: self._fim.setProx(ad) self._fim=ad def remove(self, element): if self._inicio==None: print ("Lista vazia") elif self._inicio.getValor()==element and self._inicio.getProx()==None: self._inicio=None self._inicio=None elif self._inicio.getValor()==element and self._inicio.getProx()!=None: self._inicio=self._inicio.getProx() else: delet=self._inicio.getProx() ant=self._inicio found=False while delet.getProx()!=None: if delet.getValor()==element: ant.setProx(delet.getProx()) found=True break else: ant=delet delet=delet.getProx() if found==False: print ("A lista nao tem esse elemento") def listar(self): if self._inicio==None: print ("Lista vazia") else: ult=self._inicio print (ult.getValor()) while ult.getProx()!=None: ult=ult.getProx() print (ult.getValor()) def index(self, elemento=None): if elemento==None: print ("Falta o valor") else: i=self._inicio cont=0 found=False while i!=None and found==False: if elemento==i.getValor(): found=True else: i=i.getProx() cont+=1 if found==True: return cont else: return -1 def indexSearch(self, index): ult=self._inicio if self.index(ult.getValor())==index: return ult.getValor() while ult.getProx()!=None: ult=ult.getProx() if self.index(ult.getValor())==index: return ult.getValor() print ("Indice fora do intervalo") def lenght(self): i=self._inicio cont=0 while i!=None: i=i.getProx() cont+=1 return cont class Pilha(Lista): def __init__(self): Lista.__init__(self) def pop(self): if self._inicio==None: print ("Lista vazia") return elif self._inicio.getProx()==None: element=self._inicio.getValor() self._inicio=None self._fim=None return element else: delet=self._inicio while delet.getProx()!=None: ant=delet delet=delet.getProx() self._fim=ant ant.setProx(None) return delet.getValor() def remove(self): if self._inicio==None: print ("Lista vazia") return elif self._inicio.getProx()==None: self._inicio=None self._fim=None else: delet=self._inicio while delet.getProx()!=None: ant=delet delet=delet.getProx() self._fim=ant ant.setProx(None) tokens=Pilha() tudo="" abre=open(sys.argv[1], "r") fecha=open(sys.argv[2],"w") for linha in abre: tokens.setInicio(None) tokens.setFim(None) termos=linha.split(" ") n=len(termos) while n>0: n-=1 try: x=int(termos[n]) tokens.add(x) except: x=termos[n] first=tokens.pop() second=tokens.pop() if x=="+": res=first+second tokens.add(res) elif x=="-": res=first-second tokens.add(res) elif x=="*": res=first*second tokens.add(res) elif x=="/": res=int(float(first)/float(second)) tokens.add(res) tudo+=str(tokens.getInicio())+"\n" fecha.write(tudo) abre.close() fecha.close()
[ "brunapvasconcelos@gmail.com" ]
brunapvasconcelos@gmail.com
2a5e7f5f38b91f42700324e71b0bfacd75169326
49c5c461c9805be68a318810e2ebb3381643ed59
/linkedlist/remove-loop.py
c5cee91e4abc8333b6e0eceee46a40a1b1d26633
[]
no_license
pkdism/must-do-coding
10a3ef756d24fec451d8ed09bfc22335635a6b13
ccff354eebd9b2434085528922eb3bc13715530e
refs/heads/master
2020-09-11T00:54:08.078087
2019-11-27T11:15:53
2019-11-27T11:15:53
221,886,105
0
0
null
null
null
null
UTF-8
Python
false
false
219
py
def removeTheLoop(head): h = set() prev = None it = head while it is not None: if it in h: prev.next = None break h.add(it) prev = it it = it.next
[ "pawan.dwivedi94@gmail.com" ]
pawan.dwivedi94@gmail.com
b235b3c02e8d648817c86ed6338d36c616fdc7b1
459cbeccdd385335808d9dc4a47b6658ff58a2b3
/mysite/main/migrations/0002_auto_20210612_1356.py
7dcaeaf881cde619ed25d0b6343ea602092db8e1
[]
no_license
maribelceli347/Aplicacion
41fc66c944bfb683818be92a248de0b804675e0e
29f9750bca50d940dc17c38df1f2f36bf510f77e
refs/heads/main
2023-05-28T13:32:17.792045
2021-06-12T20:02:13
2021-06-12T20:02:13
376,374,524
0
0
null
null
null
null
UTF-8
Python
false
false
775
py
# Generated by Django 3.2.4 on 2021-06-12 18:56 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('main', '0001_initial'), ] operations = [ migrations.CreateModel( name='Veterinaria', fields=[ ('id', models.BigAutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('veterinaria_nombre', models.CharField(max_length=200)), ('veterinaria_direcion', models.CharField(max_length=200)), ('veterinaria_creacion', models.DateTimeField(verbose_name='fecha de Publicacion')), ], ), migrations.DeleteModel( name='Verterinaria', ), ]
[ "79048421+maribelceli347@users.noreply.github.com" ]
79048421+maribelceli347@users.noreply.github.com
0c15c99b6035a3a37d9d9d85d732faa65f2a1760
a34a03ab7b79e772b62f344f6b3a52a784040a96
/bin/kanshin
dbe71370e51dcc616f8ca4276152a61796e9a4e0
[ "MIT" ]
permissive
basuke/kanshin-export
a413c1bb820cbdb861fb280dd66ff384075dc172
0235ce7ab570a94328524e63061830b0aa3ace74
refs/heads/master
2020-05-21T18:04:07.135221
2016-10-31T15:19:58
2016-10-31T15:19:58
64,376,495
0
0
null
null
null
null
UTF-8
Python
false
false
882
#!/usr/bin/env python import os.path import sys sys.path.insert(0, os.path.abspath('.')) import click from click import echo import kanshin.data class NativeParamType(click.ParamType): name = 'native' NATIVE = NativeParamType() # ---------------------------------------------------------- @click.group() def cli(): """CLI for accessing kanshin data""" pass @cli.group(invoke_without_command=True) @click.pass_context @click.argument('user_id', type=click.INT) def user(ctx, user_id): """Simple program that greets NAME for a total of COUNT times.""" user = kanshin.data.fetch_user(user_id) if user: if ctx.invoked_subcommand is None: echo(user) else: ctx.user = user ctx.user_id = user_id else: echo('cannot find user with id({})'.format(user_id)) if __name__ == '__main__': cli()
[ "basuke@kinoma.com" ]
basuke@kinoma.com
467cb510aa53960af2c189d8ceceb2e19665ba7a
c315e2623c0759afeb5e0390794e04b848e6df81
/fast_lemon_api.py
74c3f362d32d79a6e9623579de089e9c9d6cb668
[ "Unlicense" ]
permissive
a6502/fast_lemon_api
19b0c1f00461a697f9f4cc3ffd934f3cfe6580c2
09a5b6eec3e84d1d006f927e502a7071a28739cc
refs/heads/main
2023-03-28T11:12:51.753765
2021-03-29T23:54:31
2021-03-29T23:54:31
352,815,336
0
0
null
null
null
null
UTF-8
Python
false
false
2,481
py
#!/usr/bin/env python3 # or run as: uvicorn fast_lemon_api:app # std python-3.9 from datetime import datetime from enum import Enum from typing import List, Optional import time import uuid # python3-pydantic from pydantic import BaseModel, PositiveFloat, PositiveInt, UUID4, ValidationError, conint, constr, validator # python3-fastapi from fastapi import FastAPI, HTTPException, status from fastapi.responses import PlainTextResponse # python3-uvicorn import uvicorn # pydantic model: class OrderSide(str, Enum): buy = 'buy' sell = 'sell' class OrderStatus(str, Enum): open = 'open', executed = 'executed', # more statuseses? class BaseOrder(BaseModel): isin: constr(min_length=12, max_length=12) limit_price: PositiveFloat side: OrderSide valid_until: datetime quantity: conint(gt=0, strict=True) class NewOrder(BaseOrder): # make 'side' case insensitive @validator('side', pre=True) def validate_side(cls, s): return s.lower() @validator('valid_until', pre=True) def validate_valid_until(cls, vu): if vu <= time.time(): raise ValueError('valid_until cannot be in the past') dt = datetime.fromtimestamp(vu) return dt class Order(BaseOrder): uuid: UUID4 status: OrderStatus class Config: json_encoders = { datetime: lambda v: int(v.timestamp()), } # orders 'database' orders = {} # now build the fastapi app app = FastAPI() @app.get("/", response_class=PlainTextResponse) def read_root(): return "Welcome to the fast-lemon-api!\n" @app.get("/orders/{order_id}", response_model=Order) def get_order(order_id: uuid.UUID, q: Optional[str] = None): #return {"order_id": order_id, "q": q} if order_id in orders: return orders[order_id] else: raise HTTPException(status_code=status.HTTP_404_NOT_FOUND, detail="No such order") @app.post("/orders/", response_model=Order, status_code=status.HTTP_201_CREATED) def put_order(new_order: NewOrder): print(repr(new_order)) order_id = uuid.uuid4() order = new_order.dict() order['uuid'] = order_id order['status'] = OrderStatus.open orders[order_id] = order return order # start uvicorn when executed: if __name__ == "__main__": uvicorn.run("fast_lemon_api:app", host="127.0.0.1", port=8000, log_level="info")
[ "wieger@Dublin.a6502.net" ]
wieger@Dublin.a6502.net
7cff65e982c2b32cab03f10e594c8aaa54be7c8a
3529ecaa44a53172094ba13498097057c8972723
/Questiondir/634.find-the-derangement-of-an-array/634.find-the-derangement-of-an-array_108022165.py
4478461c59f6bddeee62b60691de01ec47975b2e
[]
no_license
cczhong11/Leetcode-contest-code-downloader
0681f0f8c9e8edd5371fd8d0a1d37dcc368566b6
db64a67869aae4f0e55e78b65a7e04f5bc2e671c
refs/heads/master
2021-09-07T15:36:38.892742
2018-02-25T04:15:17
2018-02-25T04:15:17
118,612,867
0
0
null
null
null
null
UTF-8
Python
false
false
442
py
class Solution(object): def findDerangement(self, n): """ :type n: int :rtype: int """ if n == 1: return 0 if n == 2: return 1 OPT = [0] * (n+1) OPT[0] = 1 OPT[1] = 0 OPT[2] = 1 OPT[3] = 2 for i in xrange(4, n+1): OPT[i] = (OPT[i-1] * (i-1) + (i - 1) * OPT[i-2]) % 1000000007 return OPT[n]
[ "tczhong24@gmail.com" ]
tczhong24@gmail.com
5fbe59973b26282635a73228a47461382edec61a
756cfeed032f7d206cdbe4409f2c052bf3e44fe1
/Tkinter/Lecture_4.py
0ce5a5c74a62fe228e5bc2c3eee2785792f67700
[]
no_license
Pyk017/Python
2e470a10f95aac8bb049edf13c6a825ceab68ea5
57fb48356921cc6766675620b32e4099ad371bc6
refs/heads/master
2021-07-05T00:16:22.365638
2020-12-27T08:25:10
2020-12-27T08:25:10
213,291,579
1
0
null
null
null
null
UTF-8
Python
false
false
453
py
from tkinter import * root = Tk() label1 = Label(root, text="Enter Your Name = ") label1.grid(row=0, column=0) # label1.pack() entry = Entry(root, width=25, borderwidth=5) entry.grid(row=0, column=1) # entry.pack() i = 0 def my_click(): global i label = Label(root, text="Hello " + entry.get()) i += 2 label.grid(row=i, column=0) button = Button(root, text="Click Me!", command=my_click) button.grid(row=1, column=0) root.mainloop()
[ "prakharkumar506978@gmail.com" ]
prakharkumar506978@gmail.com
0cefd1e9333d717207e4845c00ae84fb4478d05e
08a68e32dc80f99a37a30ddbbf943337546cc3d5
/.history/count/views_20200419210055.py
c0e0fbfb2b96a0104e0cd52bbfbce5cc12136149
[]
no_license
Space20001/word-count-project
dff1b4b44d2f7230070eef0d95dd968b655d92f7
795b5e8ad5c59109e96bf7a8e9192efaefa7770e
refs/heads/master
2022-04-20T17:54:05.511449
2020-04-20T15:25:46
2020-04-20T15:25:46
257,327,368
0
0
null
null
null
null
UTF-8
Python
false
false
583
py
from django.shortcuts import render # Create your views here. def home(request): return render(request, 'count/home.html', {}) def counted(request): fulltext = request.GET['fulltext'] wordlist = fulltext.split() for word in wordlist: worddictionary for word in wordlist: if word in worddictionary: else: return render(request, 'count/counted.html', {'fulltext': fulltext, 'count': len(wordlist)}) def about(request): return render(request, 'count/about.html', {about: 'about'})
[ "steve.h@blueyonder.com" ]
steve.h@blueyonder.com
7f9dd1f1f7a9135eb1ac3be360d855691bec346d
e45f1f9f3777d625c7da773f8e55589ded528711
/pic_scrapy/pic/spiders/PhotoSpider.py
bfbdf2f3503102ed8587cee37b600d0b0a2d4301
[]
no_license
631068264/learn_crawler
da973d758001b52c61aa0bb4dfc78b59a88304e4
65ac2cef7b42b0dce5fb726daa819a6ebc8ffafa
refs/heads/master
2022-11-15T13:06:55.765849
2022-10-31T07:07:31
2022-10-31T07:07:31
77,364,616
0
0
null
null
null
null
UTF-8
Python
false
false
2,504
py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ @author = 'wyx' @time = 16/10/24 15:01 @annotation = '' """ from urlparse import urljoin import scrapy from pic.items import PhotoItem class PhotoSpider(scrapy.Spider): start_urls = ["https://www.610hu.com/htm/girl.htm"] name = "photo" domain = "https://www.610hu.com" def parse(self, response): tds = response.css("table td") for td in tds: href = urljoin(self.domain, td.xpath("a/@href").extract_first()) dic = td.css("img").xpath("@src").re_first(r".*/(.*?)\.gif") yield scrapy.Request(href, callback=self.parse_page, meta={"photo_name": dic}) def parse_page(self, response): page_num = response.css(".pages strong").xpath("text()").re_first(r"/(\d?)") if page_num: for page in page_num: yield scrapy.Request(urljoin(response.url, ("%s.htm" % page)), callback=self.parse_charter, meta={"photo_name": response.meta["photo_name"]}) def parse_charter(self, response): lis = response.css("ul.movieList li") links = [] for li in lis: charter_link = urljoin(self.domain, li.xpath("a/@href").extract_first()) charter_name = li.css("h3").xpath("text()").extract_first() charter_time = li.css("span").xpath("text()").extract_first() links.append(scrapy.Request(charter_link, callback=self.parse_detail, meta={ "photo_name": response.meta["photo_name"], "charter_link": charter_link, "charter_name": charter_name, "charter_time": charter_time, })) return links def parse_detail(self, response): imgs = response.css(".picContent img") items = [] for img in imgs: src = img.xpath("@src").extract_first() item = PhotoItem({ "photo_name": response.meta["photo_name"], "charter_link": response.meta["charter_link"], "charter_name": response.meta["charter_name"], "charter_time": response.meta["charter_time"], "img_url": src, }) items.append(item) return items
[ "wyx@wuyuxideMacBook-Pro.local" ]
wyx@wuyuxideMacBook-Pro.local
0668482cec5cf3ee1d9fbb3ba9466c1dc9c9b89b
3ca67ed4d6ee727023e9342aca1fb6b2b03fac04
/coremlconverter.py
dfc9311c5cbf31381724f02600610e76b09a0d08
[]
no_license
poommomo/pokemon-classifier
17712679674361a85c5c0d986110820b11adee84
7b6724123bc2936399aed811b22c12c6d8217766
refs/heads/master
2020-03-29T04:43:43.379612
2018-08-09T08:53:12
2018-08-09T08:53:12
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,651
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- """ Created on Sun May 13 13:35:16 2018 @author: KaranJaisingh """ # import necessary packages from keras.models import load_model import coremltools import argparse import pickle # construct the argument parser and parse the arguments ap = argparse.ArgumentParser() ap.add_argument("-m", "--model", required=True, help="path to trained model model") ap.add_argument("-l", "--labelbin", required=True, help="path to label binarizer") args = vars(ap.parse_args()) # arguments passed: #1. --model: path to model #2. --labelbin: path to class label binarizer # load the class labels print("[INFO] loading class labels from label binarizer") lb = pickle.loads(open(args["labelbin"], "rb").read()) class_labels = lb.classes_.tolist() print("[INFO] class labels: {}".format(class_labels)) # load the trained convolutional neural network print("[INFO] loading model...") model = load_model(args["model"]) # convert the model to coreml format print("[INFO] converting model") coreml_model = coremltools.converters.keras.convert(model, input_names="image", image_input_names="image", image_scale=1/255.0, # very important parameter class_labels=class_labels, # obtained from LabelBinarizer object is_bgr=True) # extremely important - must be set to true is images trained with BGR colours # save the model to disk output = args["model"].rsplit(".", 1)[0] + ".mlmodel" # change the extension of model print("[INFO] saving model as {}".format(output)) coreml_model.save(output) # To run the script in Terminal: # python coremlconverter.py --model <MODEL_NAME>.model --labelbin lb.pickle
[ "KaranJaisingh@karan-jaisingh.lan" ]
KaranJaisingh@karan-jaisingh.lan
8f3c48e2d207660e14f0af89a3b6c1e6fa76b6dc
a6d9710e312caf4ae96b43f0290f9135bffdf8e0
/Unit 45/45.4.1_4/calcpkg/geometry/vector.py
6c6892b72e03d1f068f29d7872f2b5b90fd9723f
[]
no_license
gilbutITbook/006936
5b245cf1c6d4580eb07344bdaa254e4615109697
b5cd6a57cdb5bb3c2ad5e3c9471a79b3fa82bc5d
refs/heads/master
2023-01-29T07:35:46.360283
2023-01-18T06:20:49
2023-01-18T06:20:49
154,229,702
19
19
null
null
null
null
UTF-8
Python
false
false
52
py
def product(a, b): pass def dot(a, b): pass
[ "user@localhost" ]
user@localhost
e71e9d4d92731adf5fc047eb7caced983c0abefb
e0ce2909a52221a0b4b6429fd6bd4f8a67826133
/draw.py
627d53957953f7e37fb4b347adf7b52cedec107e
[]
no_license
ecrystale/mks66-solids
8c7170f5f887d8b5081fa5a57985fd91694e46df
006f9e4618a342118698c74945bbf6edb61a95f8
refs/heads/master
2020-05-07T15:41:17.966936
2019-04-18T02:31:09
2019-04-18T02:31:09
180,648,426
0
0
null
null
null
null
UTF-8
Python
false
false
11,071
py
from display import * from matrix import * from gmath import * import random def scanline_convert(polygons, i, screen, zbuffer ): color=[random.randint(0,255),random.randint(0,255),random.randint(0,255)] points=polygons[i:i+3] points.sort(key=lambda x:x[1]) bot=points[0] mid=points[1] top=points[2] x0=bot[0] x1=bot[0] z0=bot[2] z1=bot[2] if (top[1]-bot[1])!=0: xc=(top[0]-bot[0])/(top[1]-bot[1]) if (mid[1]-bot[1])!=0: xc2=(mid[0]-bot[0])/(mid[1]-bot[1]) if (top[1]-bot[1])!=0: zc1=(top[2]-bot[2])/(top[1]-bot[1]) if (mid[1]-bot[1])!=0: zc2=(mid[2]-bot[2])/(mid[1]-bot[1]) y=bot[1] while y<mid[1]: draw_line(int(x0),int(y),int(z0),int(x1),int(y),int(z1),screen,zbuffer,color) x0+=xc x1+=xc2 z0+=zc1 z1+=zc2 y+=1 x2=mid[0] z2=mid[0] #xc=(top[0]-bot[0])/(top[1]-bot[1]) xc2=0 if (top[1]-mid[1])!=0: xc2=(top[0]-mid[0])/(top[1]-mid[1]) #zc1=(top[2]-bot[2])/(top[1]-bot[1]) if (top[1]-mid[1])!=0: zc2=(top[2]-mid[2])/(top[1]-mid[1]) #y=bot[2] while y<top[1]: draw_line(int(x0),int(y),int(z0),int(x2),int(y),int(z2),screen,zbuffer,color) x0+=xc x2+=xc2 z0+=zc1 z2+=zc2 y+=1 def add_polygon( polygons, x0, y0, z0, x1, y1, z1, x2, y2, z2 ): add_point(polygons, x0, y0, z0) add_point(polygons, x1, y1, z1) add_point(polygons, x2, y2, z2) def draw_polygons( polygons, screen, zbuffer, color ): if len(polygons) < 2: print ('Need at least 3 points to draw') return point = 0 while point < len(polygons) - 2: normal = calculate_normal(polygons, point)[:] #print normal if normal[2] > 0: scanline_convert(polygons,point,screen,zbuffer) point+= 3 ''' draw_line( int(polygons[point][0]), int(polygons[point][1]), polygons[point][2], int(polygons[point+1][0]), int(polygons[point+1][1]), polygons[point+1][2], screen, zbuffer, color) draw_line( int(polygons[point+2][0]), int(polygons[point+2][1]), polygons[point+2][2], int(polygons[point+1][0]), int(polygons[point+1][1]), polygons[point+1][2], screen, zbuffer, color) draw_line( int(polygons[point][0]), int(polygons[point][1]), polygons[point][2], int(polygons[point+2][0]), int(polygons[point+2][1]), polygons[point+2][2], screen, zbuffer, color)''' def add_box( polygons, x, y, z, width, height, depth ): x1 = x + width y1 = y - height z1 = z - depth #front add_polygon(polygons, x, y, z, x1, y1, z, x1, y, z) add_polygon(polygons, x, y, z, x, y1, z, x1, y1, z) #back add_polygon(polygons, x1, y, z1, x, y1, z1, x, y, z1) add_polygon(polygons, x1, y, z1, x1, y1, z1, x, y1, z1) #right side add_polygon(polygons, x1, y, z, x1, y1, z1, x1, y, z1) add_polygon(polygons, x1, y, z, x1, y1, z, x1, y1, z1) #left side add_polygon(polygons, x, y, z1, x, y1, z, x, y, z) add_polygon(polygons, x, y, z1, x, y1, z1, x, y1, z) #top add_polygon(polygons, x, y, z1, x1, y, z, x1, y, z1) add_polygon(polygons, x, y, z1, x, y, z, x1, y, z) #bottom add_polygon(polygons, x, y1, z, x1, y1, z1, x1, y1, z) add_polygon(polygons, x, y1, z, x, y1, z1, x1, y1, z1) def add_sphere(polygons, cx, cy, cz, r, step ): points = generate_sphere(cx, cy, cz, r, step) lat_start = 0 lat_stop = step longt_start = 0 longt_stop = step step+= 1 for lat in range(lat_start, lat_stop): for longt in range(longt_start, longt_stop): p0 = lat * step + longt p1 = p0+1 p2 = (p1+step) % (step * (step-1)) p3 = (p0+step) % (step * (step-1)) if longt != step - 2: add_polygon( polygons, points[p0][0], points[p0][1], points[p0][2], points[p1][0], points[p1][1], points[p1][2], points[p2][0], points[p2][1], points[p2][2]) if longt != 0: add_polygon( polygons, points[p0][0], points[p0][1], points[p0][2], points[p2][0], points[p2][1], points[p2][2], points[p3][0], points[p3][1], points[p3][2]) def generate_sphere( cx, cy, cz, r, step ): points = [] rot_start = 0 rot_stop = step circ_start = 0 circ_stop = step for rotation in range(rot_start, rot_stop): rot = rotation/float(step) for circle in range(circ_start, circ_stop+1): circ = circle/float(step) x = r * math.cos(math.pi * circ) + cx y = r * math.sin(math.pi * circ) * math.cos(2*math.pi * rot) + cy z = r * math.sin(math.pi * circ) * math.sin(2*math.pi * rot) + cz points.append([x, y, z]) #print 'rotation: %d\tcircle%d'%(rotation, circle) return points def add_torus(polygons, cx, cy, cz, r0, r1, step ): points = generate_torus(cx, cy, cz, r0, r1, step) lat_start = 0 lat_stop = step longt_start = 0 longt_stop = step for lat in range(lat_start, lat_stop): for longt in range(longt_start, longt_stop): p0 = lat * step + longt; if (longt == (step - 1)): p1 = p0 - longt; else: p1 = p0 + 1; p2 = (p1 + step) % (step * step); p3 = (p0 + step) % (step * step); add_polygon(polygons, points[p0][0], points[p0][1], points[p0][2], points[p3][0], points[p3][1], points[p3][2], points[p2][0], points[p2][1], points[p2][2] ) add_polygon(polygons, points[p0][0], points[p0][1], points[p0][2], points[p2][0], points[p2][1], points[p2][2], points[p1][0], points[p1][1], points[p1][2] ) def generate_torus( cx, cy, cz, r0, r1, step ): points = [] rot_start = 0 rot_stop = step circ_start = 0 circ_stop = step for rotation in range(rot_start, rot_stop): rot = rotation/float(step) for circle in range(circ_start, circ_stop): circ = circle/float(step) x = math.cos(2*math.pi * rot) * (r0 * math.cos(2*math.pi * circ) + r1) + cx; y = r0 * math.sin(2*math.pi * circ) + cy; z = -1*math.sin(2*math.pi * rot) * (r0 * math.cos(2*math.pi * circ) + r1) + cz; points.append([x, y, z]) return points def add_circle( points, cx, cy, cz, r, step ): x0 = r + cx y0 = cy i = 1 while i <= step: t = float(i)/step x1 = r * math.cos(2*math.pi * t) + cx; y1 = r * math.sin(2*math.pi * t) + cy; add_edge(points, x0, y0, cz, x1, y1, cz) x0 = x1 y0 = y1 i+= 1 def add_curve( points, x0, y0, x1, y1, x2, y2, x3, y3, step, curve_type ): xcoefs = generate_curve_coefs(x0, x1, x2, x3, curve_type)[0] ycoefs = generate_curve_coefs(y0, y1, y2, y3, curve_type)[0] i = 1 while i <= step: t = float(i)/step x = t * (t * (xcoefs[0] * t + xcoefs[1]) + xcoefs[2]) + xcoefs[3] y = t * (t * (ycoefs[0] * t + ycoefs[1]) + ycoefs[2]) + ycoefs[3] #x = xcoefs[0] * t*t*t + xcoefs[1] * t*t + xcoefs[2] * t + xcoefs[3] #y = ycoefs[0] * t*t*t + ycoefs[1] * t*t + ycoefs[2] * t + ycoefs[3] add_edge(points, x0, y0, 0, x, y, 0) x0 = x y0 = y i+= 1 def draw_lines( matrix, screen, zbuffer, color ): if len(matrix) < 2: print ('Need at least 2 points to draw') return point = 0 while point < len(matrix) - 1: draw_line( int(matrix[point][0]), int(matrix[point][1]), matrix[point][2], int(matrix[point+1][0]), int(matrix[point+1][1]), matrix[point+1][2], screen, zbuffer, color) point+= 2 def add_edge( matrix, x0, y0, z0, x1, y1, z1 ): add_point(matrix, x0, y0, z0) add_point(matrix, x1, y1, z1) def add_point( matrix, x, y, z=0 ): matrix.append( [x, y, z, 1] ) def draw_line( x0, y0, z0, x1, y1, z1, screen, zbuffer, color ): #swap points if going right -> left if x0 > x1: xt = x0 yt = y0 x0 = x1 y0 = y1 x1 = xt y1 = yt zt = z0 z0 = z1 z1 = zt x = x0 y = y0 z = z0 A = 2 * (y1 - y0) B = -2 * (x1 - x0) wide = False tall = False if ( abs(x1-x0) >= abs(y1 - y0) ): #octants 1/8 wide = True loop_start = x loop_end = x1 dx_east = dx_northeast = 1 dy_east = 0 d_east = A if ( A > 0 ): #octant 1 d = A + B/2 dy_northeast = 1 d_northeast = A + B else: #octant 8 d = A - B/2 dy_northeast = -1 d_northeast = A - B else: #octants 2/7 tall = True dx_east = 0 dx_northeast = 1 if ( A > 0 ): #octant 2 d = A/2 + B dy_east = dy_northeast = 1 d_northeast = A + B d_east = B loop_start = y loop_end = y1 else: #octant 7 d = A/2 - B dy_east = dy_northeast = -1 d_northeast = A - B d_east = -1 * B loop_start = y1 loop_end = y while ( loop_start < loop_end ): plot( screen, zbuffer, color, x, y, z0) if ( (wide and ((A > 0 and d > 0) or (A < 0 and d < 0))) or (tall and ((A > 0 and d < 0) or (A < 0 and d > 0 )))): x+= dx_northeast y+= dy_northeast d+= d_northeast else: x+= dx_east y+= dy_east d+= d_east loop_start+= 1 plot( screen, zbuffer, color, x, y, z0)
[ "elee19@stuy.edu" ]
elee19@stuy.edu