blob_id
stringlengths
40
40
directory_id
stringlengths
40
40
path
stringlengths
3
281
content_id
stringlengths
40
40
detected_licenses
listlengths
0
57
license_type
stringclasses
2 values
repo_name
stringlengths
6
116
snapshot_id
stringlengths
40
40
revision_id
stringlengths
40
40
branch_name
stringclasses
313 values
visit_date
timestamp[us]
revision_date
timestamp[us]
committer_date
timestamp[us]
github_id
int64
18.2k
668M
star_events_count
int64
0
102k
fork_events_count
int64
0
38.2k
gha_license_id
stringclasses
17 values
gha_event_created_at
timestamp[us]
gha_created_at
timestamp[us]
gha_language
stringclasses
107 values
src_encoding
stringclasses
20 values
language
stringclasses
1 value
is_vendor
bool
2 classes
is_generated
bool
2 classes
length_bytes
int64
4
6.02M
extension
stringclasses
78 values
content
stringlengths
2
6.02M
authors
listlengths
1
1
author
stringlengths
0
175
0dfddc24c14a9eec82fe2903f8bad926bc3e307c
262cd372c53fbc002bd0966129e6539a0474b7c9
/award/tests.py
a22138b6ee738a458720e4740139623cb1e2887a
[ "MIT" ]
permissive
owinolawrence/awwards
e83bd2f8a3e8db2d0a3a5c2316d4f6db5a8acf08
23ec5a53e715c3a75139c31a8d340c7b348e57a8
refs/heads/master
2021-09-10T01:00:36.329048
2020-01-16T11:39:01
2020-01-16T11:39:01
233,040,878
0
0
MIT
2021-09-08T01:35:38
2020-01-10T12:18:52
Python
UTF-8
Python
false
false
807
py
from django.test import TestCase from .models import * import unittest from django.contrib.auth.models import User # Create your tests here. class TestProfile(TestCase): def setUp(self): self.post = Post(title='title',live_link='sfggfdsgedgfgfd',description ='sdfdsad',country='dasfsdfasdf',languages='sfsafsasfa',landing_page = 'media/img.jpeg',screenshot_one = 'media/img.jpeg',screenshot_two = 'media/img.jpeg',screenshot_three = 'media/img.jpeg',screenshot_four = 'media/img.jpeg',date_posted='hhhjjhjhh') def test_instance(self): self.assertTrue(isinstance(self.post, Post)) class TestReview(TestCase): def setUp(self): self.review = Review(design=1,usability=1,content=1) def test_instance(self): self.assertTrue(isinstance(self.review,Review))
[ "owinolawrence21@gmail.com" ]
owinolawrence21@gmail.com
eb8412cedc39fa28a1baad4fe8b90e486f742869
ea6cf2360eead61c3534beeb1f394d0fe0905ee9
/myapp/migrations/0001_initial.py
6f78a0420ad000f89a3642691ef23eb52793f9e8
[]
no_license
123tian/mypro
9f5d89cdf13945525ea943fbe38f47a8c5d52219
e0c1d5691130139f16359fdf0099e8d12c310eb9
refs/heads/master
2020-07-13T15:22:49.251644
2019-08-30T00:50:11
2019-08-30T00:50:11
205,105,716
0
0
null
null
null
null
UTF-8
Python
false
false
849
py
# -*- coding: utf-8 -*- # Generated by Django 1.11.22 on 2019-08-28 07:53 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='Publisher', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=32, unique=True, verbose_name='名称')), ('adddress', models.CharField(max_length=128, verbose_name='地址')), ], options={ 'db_table': 'pub_db', 'verbose_name_plural': '出版社', 'verbose_name': '出版社', }, ), ]
[ "3053366571@qq.com" ]
3053366571@qq.com
7771cd92728727fe710936fbd0651fe13a42bb3c
d0a75ad618998a34fb22b650697b171b25c8a724
/linkchecker/settings.py
bc7a8a0282c7cd5f992376d3c9a7c4f3aa3e07a0
[ "Unlicense" ]
permissive
kohout/django-linkchecker
05f0b853cae4c6a35e6b305411ef4d2243f2dea5
34cf8b0a4a172b22dad525329ee0bfd4eca8136b
refs/heads/master
2021-01-21T13:49:44.605414
2016-05-17T14:52:05
2016-05-17T14:52:05
45,972,769
1
0
null
null
null
null
UTF-8
Python
false
false
992
py
# -*- coding: utf-8 -*- from django.conf import settings from django.db.models import URLField import logging def get_default_logger(): logger = logging.getLogger('link_checker') logger.setLevel(logging.INFO) formatter = logging.Formatter('%(asctime)s - %(levelname)s: %(message)s') console_handler = logging.StreamHandler() console_handler.setFormatter(formatter) logger.addHandler(console_handler) return logger LINKCHECKER_DELETE_HISTORY_ON_UPDATE = getattr(settings, 'LINKCHECKER_DELETE_HISTORY_ON_UPDATE', False) LINKCHECKER_LOGGER = getattr(settings, 'LINKCHECKER_LOGGER', get_default_logger()) LINKCHECKERS = getattr(settings, 'LINKCHECKERS', list()) LINKCHECKER_FIELD_TYPES = getattr(settings, 'LINKCHECKER_FIELD_TYPES', [URLField]) LINKCHECKER_LINK_TO = getattr(settings, 'LINKCHECKER_LINK_TO', None) LINKCHECKER_TIMEOUT = getattr(settings, 'LINKCHECKER_TIMEOUT', 3) LINKCHECKER_IGNORE_TIMEOUT = getattr(settings, 'LINKCHECKER_IGNORE_TIMEOUT', False)
[ "dm@getaweb.at" ]
dm@getaweb.at
ad4bbd701c27d99b9d0dac95c62c0fc2e8227f82
4bfa2ea9963a9832430720fbbdef459ca8b5cd94
/checkSolution.py
185f735060db50c125e778b5514e35f2fbe099a6
[]
no_license
franciscolemos/csp
5d0df21c029f54a7c72e339c89b1fc58ef9bf20f
7fd26b6eb3841b9e56bc86dafcccb0c8a1cadf01
refs/heads/master
2023-05-07T18:03:56.480043
2021-05-19T23:47:19
2021-05-19T23:47:19
231,985,455
0
0
null
null
null
null
UTF-8
Python
false
false
2,552
py
from recovery.repositories.paths import pathList import subprocess import os from datetime import datetime import pandas as pd from recovery.dal import resultsDtype import numpy as np sizeLine = {} sizeLine['version'] = [] sizeLine['dateTime'] = [] sizeLine['dataInstance'] = [] sizeLine['checkAircraftBreakdownPeriod'] = [] sizeLine['checkAircraftCapacity'] = [] sizeLine['checkAircraftCreation'] = [] sizeLine['checkAircraftSwap'] = [] sizeLine['checkAirportCapacity'] = [] sizeLine['checkCancellationofCreatedRotation'] = [] sizeLine['checkFixedFlights'] = [] sizeLine['checkFlight'] = [] sizeLine['checkItinerary'] = [] sizeLine['checkPassengerReac'] = [] sizeLine['checkRotation'] = [] sizeLine['costs'] = [] today = datetime.now().strftime("%Y%m%d_%H%M") for path in pathList: dataSet = path.split("/")[-1] print(dataSet) os.chdir(path) os.system("solutionChecker-win32.exe") sizeLine['version'] = '32/32 lb 3**12' sizeLine['dateTime'].append(today) sizeLine['dataInstance'].append(dataSet) sizeLine['checkAircraftBreakdownPeriod'].append(os.path.getsize("./results/checkAircraftBreakdownPeriod.txt")) sizeLine['checkAircraftCapacity'].append(os.path.getsize("./results/checkAircraftCapacity.txt")) sizeLine['checkAircraftCreation'].append(os.path.getsize("./results/checkAircraftCreation.txt")) sizeLine['checkAircraftSwap'].append(os.path.getsize("./results/checkAircraftSwap.txt")) sizeLine['checkAirportCapacity'].append(os.path.getsize("./results/checkAirportCapacity.txt")) sizeLine['checkCancellationofCreatedRotation'].append(os.path.getsize("./results/checkCancellationofCreatedRotation.txt")) sizeLine['checkFixedFlights'].append(os.path.getsize("./results/checkFixedFlights.txt")) sizeLine['checkFlight'].append(os.path.getsize("./results/checkFlight.txt")) sizeLine['checkItinerary'].append(os.path.getsize("./results/checkItinerary.txt")) sizeLine['checkPassengerReac'].append(os.path.getsize("./results/checkPassengerReac.txt")) sizeLine['checkRotation'].append(os.path.getsize("./results/checkRotation.txt")) os.system("costChecker-win32.exe > results/cost.txt") costs = np.genfromtxt('results/cost.txt', delimiter=':') sizeLine['costs'].append(costs[-1][1]) os.chdir("../../../") dataSetSizes = pd.DataFrame(sizeLine) try: dataSetSizes.to_csv('./results/prpInf.csv', mode='a', index=False, header = False) except: import pdb; pdb.set_trace() dataSetSizes.to_csv('./results/prpInf.csv', mode='a', index=False, header = False)
[ "chico.lemos@gmail.com" ]
chico.lemos@gmail.com
62aa187e75d0d640cc2d69a224a102a0cafca5fc
0179a8c11f51d89cc962c7d9249203ff0e67e405
/shell/shell_contract.py
1238844b436bb74f76efe8dda6e11fca6a4f0c77
[ "MIT" ]
permissive
paulo-romano/orcamentos
7033637065c39c457a59b53eab215234f7d5b85a
dc87fd2736e9f8262ed775bf9160d1e21eb1684a
refs/heads/master
2021-01-15T22:14:41.595934
2016-02-16T23:02:59
2016-02-16T23:02:59
51,551,215
1
0
null
2016-02-11T22:16:43
2016-02-11T22:16:43
null
UTF-8
Python
false
false
523
py
from random import choice from django.db import IntegrityError from orcamentos.core.models import Contract, Proposal, Customer REPEAT = Proposal.objects.filter(status='a') for i in REPEAT: proposal = Proposal.objects.get(pk=i.pk) contractor = Customer.objects.get(pk=proposal.work.customer.pk) try: Contract.objects.create( proposal=proposal, contractor=contractor, is_canceled=choice((True, False))) except IntegrityError: print('Registro existente.')
[ "rg3915@yahoo.com.br" ]
rg3915@yahoo.com.br
b42de3203dd191db3af92bba82716a215ca39369
e32297163bf143b194e438783f93c61c5dc08e11
/venv/Scripts/pip3.8-script.py
9256ee434aaad8237ff00414d9a033378a1b7472
[]
no_license
eonnyhoney/animal
be185f65eeff4964c0e311ae10b761703f479c67
8cbf560b83509446d652fc0d59c2051278953653
refs/heads/master
2020-12-30T07:51:01.351800
2020-02-07T12:18:38
2020-02-07T12:18:38
238,914,622
0
0
null
null
null
null
WINDOWS-1252
Python
false
false
427
py
#!C:\Users\Eon2\Desktop\½ºÆÄ¸£Å¸\animal\venv\Scripts\python.exe -x # EASY-INSTALL-ENTRY-SCRIPT: 'pip==19.0.3','console_scripts','pip3.8' __requires__ = 'pip==19.0.3' import re import sys from pkg_resources import load_entry_point if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit( load_entry_point('pip==19.0.3', 'console_scripts', 'pip3.8')() )
[ "59526689+eonnyhoney@users.noreply.github.com" ]
59526689+eonnyhoney@users.noreply.github.com
9cf98b7b4745bf18117c0e68108e370d4226cd25
24e21c68bc2c4f1c3f58b96ae13512968a919024
/memoryAndMulti/threadDemo.py
557b1eb374bb59d12ee08ff31de2c68f27abdcf2
[]
no_license
maketubu7/spiderDemo
0308e88815c2035fa33acd1c4ca85329d2435034
9c5e78fdafba37a08e51c2e988c54957feed5b0f
refs/heads/master
2021-02-09T18:43:32.493539
2020-11-11T09:13:21
2020-11-11T09:13:21
244,314,824
0
0
null
null
null
null
UTF-8
Python
false
false
899
py
# -*- coding: utf-8 -*- # @Time : 2020/2/27 0:10 # @Author : Deng Wenxing # @Email : dengwenxingae86@163.com # @File : threadDemo.py # @Software: PyCharm from threading import Thread import threading,time from typing import Optional def loop(): print(threading.current_thread().name) n = 0 while n < 5: print(n) n += 1 def use_thread(): print(threading.current_thread().name) t = Thread(target=loop,name='loop_thread') ##启动 t.start() ##挂起 t.join() class my_thread(Thread): def __init__(self): super(my_thread,self).__init__() self.n = 0 def run(self): while self.n < 5: print(self.n) print(threading.current_thread().name) time.sleep(1) self.n += 1 if __name__ == "__main__": # use_thread() t = my_thread() t.start() t.join()
[ "601176930@qq.com" ]
601176930@qq.com
ad5de07133769eec2ccac7a11232619b4444be85
d9e55c4afe0db76c39d992374a9a62c4fb81d473
/tests/test_features/test_tile_coding.py
0bb259f29a8f91fd41a898ef46295c7718161019
[ "BSD-3-Clause" ]
permissive
rldotai/flib
36da0f2cd9d2220b42dc788a1a0c3c90359762e5
695e875f708b0b71c9b005fdf85c066e4ffb7c0f
refs/heads/master
2021-01-10T21:26:38.225855
2015-06-14T01:06:01
2015-06-14T01:06:01
35,320,396
1
0
null
null
null
null
UTF-8
Python
false
false
1,927
py
""" Tests for tile_coding.py """ import pytest import numpy as np import flib from flib import TileCoder def test_init(): n_input = 4 n_output = 16 n_tiles = 1000 f = TileCoder(n_input, n_output, n_tiles) f = TileCoder(n_input, n_output, n_tiles, scale=np.arange(n_input)) f = TileCoder(n_input, n_output, n_tiles, table_size=512) f = TileCoder(n_input, n_output, n_tiles, random_seed=123) def test_output_range(): n_input = 4 n_output = 16 n_tiles = 1000 f = TileCoder(n_input, n_output, n_tiles) low, high = 0, 100 inputs = np.random.uniform(low, high, size=(1000, n_input)) for i in inputs: out = f(i) assert(np.all(0 <= out)) assert(np.all(n_tiles > out)) def test_output_shape(): cases = 1000 n_input = 4 n_output = 16 n_tiles = 1000 f = TileCoder(n_input, n_output, n_tiles) low, high = 0, 100 inputs = np.random.uniform(low, high, size=(cases, n_input)) # Test individual outputs for i in inputs: out = f(i) assert(len(out) == n_output) # Test coding for multiple inputs at the same time outputs = f(inputs) assert(outputs.shape == (cases, n_output)) def test_randomization(): cases = 10000 n_input = 4 n_output = 16 n_tiles = 1000 f = TileCoder(n_input, n_output, n_tiles) low, high = 0, 100 inputs = np.random.uniform(low, high, size=(cases, n_input)) outputs = f(inputs) # Attempt to quantify how random the tile coding was tiles = outputs.flatten() total_tiles = len(tiles) nbins = 50 hist, bins = np.histogram(tiles, bins=np.linspace(0, n_tiles, nbins)) # More than three standard deviations away is a cause for worry zscore = np.abs(hist - hist.mean())/np.std(hist) assert(np.all(zscore < 3)) # TODO: test that varying a single element of the input causes appropriate change in output
[ "rldot41@gmail.com" ]
rldot41@gmail.com
294ccce0e1d57d9c848ae8d1b3031470469241bc
1f22a46c06755e518907866061dd17c3285374e9
/com_sqlite/common.py
ca9d65d2e0e23b86d0edf7d3a0c37f5e37b3f414
[]
no_license
zhng1200/170731
8a53fecc6e6cd830e036ab4448768f6a459acfdc
29bf71bce8de062b24717f41f3a55bb7b1f5486c
refs/heads/master
2021-06-24T19:24:07.062974
2017-09-18T07:00:53
2017-09-18T07:00:53
103,904,024
0
0
null
null
null
null
UTF-8
Python
false
false
464
py
# coding=utf-8 import sqlite3 class SQLiteDataBase: def __init__(self): self.db = sqlite3.connect("mysite.db") self.cursor = self.db.cursor() def query(self, sql): self.cursor.execute(sql) self.db.commit() results = self.cursor.fetchall() return results def execute(self, sql): self.cursor.execute(sql) self.db.commit() return 1 def __del__(self): self.db.close()
[ "zhaoning_zb@cmos.chinamobile.com" ]
zhaoning_zb@cmos.chinamobile.com
add30295e413442fa81dfc795c0befa9db64f85a
61ce05b0b62e0ab7ccd0ca90958268f6ac18127a
/Chapters Code/ch3_Board_Representation_and_Rules/dlgo/goboard_slow.py
6e0d3a4e168e4388e9dfc8268a4ee58ee0f53e6a
[]
no_license
HagarHaytham/Ghost-GO
0d7c90b08da9a20f725229b2852ca91df2df44be
f23c416c08d44b83469b09c5c55031fa33d9ef8f
refs/heads/master
2022-03-26T09:48:01.102320
2019-12-16T05:29:09
2019-12-16T05:29:09
215,604,238
1
0
null
null
null
null
UTF-8
Python
false
false
8,855
py
import numpy as np # tag::imports[] import copy from dlgo.gotypes import Player # end::imports[] from dlgo.gotypes import Point from dlgo.scoring import compute_game_result __all__ = [ 'Board', 'GameState', 'Move', ] class IllegalMoveError(Exception): pass # tag::strings[] class GoString(): # <1> def __init__(self, color, stones, liberties): self.color = color self.stones = set(stones) self.liberties = set(liberties) def remove_liberty(self, point): self.liberties.remove(point) def add_liberty(self, point): self.liberties.add(point) def merged_with(self, go_string): # <2> assert go_string.color == self.color combined_stones = self.stones | go_string.stones return GoString( self.color, combined_stones, (self.liberties | go_string.liberties) - combined_stones) @property def num_liberties(self): return len(self.liberties) def __eq__(self, other): return isinstance(other, GoString) and \ self.color == other.color and \ self.stones == other.stones and \ self.liberties == other.liberties # <1> Go strings are stones that are linked by a chain of connected stones of the same color. # <2> Return a new Go string containing all stones in both strings. # end::strings[] # tag::board_init[] class Board(): # <1> def __init__(self, num_rows, num_cols): self.num_rows = num_rows self.num_cols = num_cols self._grid = {} # <1> A board is initialized as empty grid with the specified number of rows and columns. # end::board_init[] # tag::board_place_0[] def place_stone(self, player, point): assert self.is_on_grid(point) assert self._grid.get(point) is None adjacent_same_color = [] adjacent_opposite_color = [] liberties = [] for neighbor in point.neighbors(): # <1> if not self.is_on_grid(neighbor): continue neighbor_string = self._grid.get(neighbor) if neighbor_string is None: liberties.append(neighbor) elif neighbor_string.color == player: if neighbor_string not in adjacent_same_color: adjacent_same_color.append(neighbor_string) else: if neighbor_string not in adjacent_opposite_color: adjacent_opposite_color.append(neighbor_string) new_string = GoString(player, [point], liberties) # <1> First, we examine direct neighbors of this point. # end::board_place_0[] # tag::board_place_1[] for same_color_string in adjacent_same_color: # <1> new_string = new_string.merged_with(same_color_string) for new_string_point in new_string.stones: self._grid[new_string_point] = new_string for other_color_string in adjacent_opposite_color: # <2> other_color_string.remove_liberty(point) for other_color_string in adjacent_opposite_color: # <3> if other_color_string.num_liberties == 0: self._remove_string(other_color_string) # <1> Merge any adjacent strings of the same color. # <2> Reduce liberties of any adjacent strings of the opposite color. # <3> If any opposite color strings now have zero liberties, remove them. # end::board_place_1[] # tag::board_remove[] def _remove_string(self, string): for point in string.stones: for neighbor in point.neighbors(): # <1> neighbor_string = self._grid.get(neighbor) if neighbor_string is None: continue if neighbor_string is not string: neighbor_string.add_liberty(point) del(self._grid[point]) # <1> Removing a string can create liberties for other strings. # end::board_remove[] # tag::board_utils[] def is_on_grid(self, point): return 1 <= point.row <= self.num_rows and \ 1 <= point.col <= self.num_cols def get(self, point): # <1> string = self._grid.get(point) if string is None: return None return string.color def get_go_string(self, point): # <2> string = self._grid.get(point) if string is None: return None return string # <1> Returns the content of a point on the board: a Player if there is a stone on that point or else None. # <2> Returns the entire string of stones at a point: a GoString if there is a stone on that point or else None. # end::board_utils[] def __eq__(self, other): return isinstance(other, Board) and \ self.num_rows == other.num_rows and \ self.num_cols == other.num_cols and \ self._grid == other._grid # tag::moves[] class Move(): # <1> def __init__(self, point=None, is_pass=False, is_resign=False): assert (point is not None) ^ is_pass ^ is_resign self.point = point self.is_play = (self.point is not None) self.is_pass = is_pass self.is_resign = is_resign @classmethod def play(cls, point): # <2> return Move(point=point) @classmethod def pass_turn(cls): # <3> return Move(is_pass=True) @classmethod def resign(cls): # <4> return Move(is_resign=True) # <1> Any action a player can play on a turn, either is_play, is_pass or is_resign will be set. # <2> This move places a stone on the board. # <3> This move passes. # <4> This move resigns the current game # end::moves[] # tag::game_state[] class GameState(): def __init__(self, board, next_player, previous, move): self.board = board self.next_player = next_player self.previous_state = previous self.last_move = move def apply_move(self, move): # <1> if move.is_play: next_board = copy.deepcopy(self.board) next_board.place_stone(self.next_player, move.point) else: next_board = self.board return GameState(next_board, self.next_player.other, self, move) @classmethod def new_game(cls, board_size): if isinstance(board_size, int): board_size = (board_size, board_size) board = Board(*board_size) return GameState(board, Player.black, None, None) # <1> Return the new GameState after applying the move. # end::game_state[] # tag::self_capture[] def is_move_self_capture(self, player, move): if not move.is_play: return False next_board = copy.deepcopy(self.board) next_board.place_stone(player, move.point) new_string = next_board.get_go_string(move.point) return new_string.num_liberties == 0 # end::self_capture[] # tag::is_ko[] @property def situation(self): return (self.next_player, self.board) def does_move_violate_ko(self, player, move): if not move.is_play: return False next_board = copy.deepcopy(self.board) next_board.place_stone(player, move.point) next_situation = (player.other, next_board) past_state = self.previous_state while past_state is not None: if past_state.situation == next_situation: return True past_state = past_state.previous_state return False # end::is_ko[] # tag::is_valid_move[] def is_valid_move(self, move): if self.is_over(): return False if move.is_pass or move.is_resign: return True return ( self.board.get(move.point) is None and not self.is_move_self_capture(self.next_player, move) and not self.does_move_violate_ko(self.next_player, move)) # end::is_valid_move[] # tag::is_over[] def is_over(self): if self.last_move is None: return False if self.last_move.is_resign: return True second_last_move = self.previous_state.last_move if second_last_move is None: return False return self.last_move.is_pass and second_last_move.is_pass # end::is_over[] def legal_moves(self): moves = [] for row in range(1, self.board.num_rows + 1): for col in range(1, self.board.num_cols + 1): move = Move.play(Point(row, col)) if self.is_valid_move(move): moves.append(move) # These two moves are always legal. moves.append(Move.pass_turn()) moves.append(Move.resign()) return moves def winner(self): if not self.is_over(): return None if self.last_move.is_resign: return self.next_player game_result = compute_game_result(self) return game_result.winner
[ "mohamedtalaat0111790@gmail.com" ]
mohamedtalaat0111790@gmail.com
fda1f90a4be88c7944f2879764d5c153faed9cb0
c57439f0c98af370ace65f9d55ef5a457bedc531
/ydk/models/ipv6/Cisco_IOS_XR_ipv6_ma_subscriber_cfg.py
a66d84f0a2924a7e9df63458243f00228eb1dd1d
[ "Apache-2.0" ]
permissive
myahmao/ydk-py
c932fbd8245e554227cce0fd723d9a22887b0c40
2f367d93f2088d4abdc2f2bb10ca4864952b458a
refs/heads/master
2021-01-14T11:32:29.064494
2016-03-15T22:44:05
2016-03-15T22:44:05
null
0
0
null
null
null
null
UTF-8
Python
false
false
928
py
""" Cisco_IOS_XR_ipv6_ma_subscriber_cfg This module contains a collection of YANG definitions for Cisco IOS\-XR ipv6\-ma\-subscriber package configuration. This YANG module augments the Cisco\-IOS\-XR\-subscriber\-infra\-tmplmgr\-cfg module with configuration data. Copyright (c) 2013\-2015 by Cisco Systems, Inc. All rights reserved. """ import re import collections from enum import Enum from ydk.types import Empty, YList, DELETE, Decimal64, FixedBitsDict from ydk.errors import YPYError, YPYDataValidationError class Ipv6ReachableVia_Enum(Enum): """ Ipv6ReachableVia_Enum Ipv6 reachable via """ """ Source is reachable via interface on which packet was received """ RECEIVED = 1 @staticmethod def _meta_info(): from ydk.models.ipv6._meta import _Cisco_IOS_XR_ipv6_ma_subscriber_cfg as meta return meta._meta_table['Ipv6ReachableVia_Enum']
[ "manradha@cisco.com" ]
manradha@cisco.com
cf8df78c19fed7972b683782a743137388fcee12
6b518cf14ea3f59fd59136dbd2a7ac70234bb96e
/pspipe.py
4523f7e32db887641957d2c80753873e9e831bcc
[]
no_license
simula67/advanced-python-course-material
8064a1adddff45b0980d4bd1948fdeb2f88aec89
98870da337cbc001bcf4215ce44f82f0430fd3ce
refs/heads/master
2016-09-06T12:29:37.397321
2015-06-29T05:10:19
2015-06-29T05:10:19
38,228,793
0
0
null
null
null
null
UTF-8
Python
false
false
289
py
__author__ = 'antonjoj' import subprocess cat = subprocess.Popen('type datafiles\\passwd', shell=True, stdout=subprocess.PIPE) find = subprocess.Popen('find \"root\"', stdout=subprocess.PIPE, shell=True, stdin=cat.stdout) for line in find.communicate(): if line: print line
[ "simula67@gmail.com" ]
simula67@gmail.com
5739e561cd3360ae20088b5f64ed45c14b854723
cfb44550355ea3c36e610d3f1eb75d8dcbdc8ebe
/strawberry/setup.py
62732a79141f77a94d9dc5da70b6e018a0858ff8
[ "Apache-2.0" ]
permissive
KiritoDv/blueberry
bd0dc38bfe16622693efd8ff4a31368b4dbeb4ac
a47feeb3e944d44b9f2af4661f6c409f51fbabd6
refs/heads/master
2023-01-18T19:06:00.074071
2020-11-24T23:04:09
2020-11-24T23:04:09
314,765,853
0
0
null
null
null
null
UTF-8
Python
false
false
239
py
from distutils.core import setup, Extension module1 = Extension('strawberry', sources = ['straw.c']) setup (name = 'strawberry', version = '1.0', description = 'Strawberry miniaudio wrapper', ext_modules = [module1])
[ "alex@pop-os.localdomain" ]
alex@pop-os.localdomain
1ef6fa14c5eddb4800ee9cb3c0f1c9d522ce8cf8
826a4e7a2d80a802186480cd88201d4d547c9239
/Assignment 5/TextProcessor.py
526088cae4c65bbc6f25d7ebdb5975d006c28853
[]
no_license
ThomasBakkenMoe/IINI4014
3b971d952ddad0ffbe6d78a4c44e8dfe2f868e36
ed43382ef1fff34c951444ea051cb6a925badc15
refs/heads/master
2022-12-18T00:26:35.085578
2020-09-21T13:02:31
2020-09-21T13:02:31
297,340,346
0
0
null
null
null
null
UTF-8
Python
false
false
3,450
py
import os import re class WordAndNumber: ''' Object that contains a word and the frequency of that word Constructor takes: word: string: the saved word frequency: int: the frequency of the saved word ''' def __init__(self, word, frequency): self.word = word self.frequency = frequency def __repr__(self): str = "{}: {}" return str.format(self.word, self.frequency) def findFile(filename): ''' Function that finds a file in the subdirectory tree beneath the python file takes: :param filename: string: the name of the file to be found :return: string: the relative path to the file ''' for root, dirs, files in os.walk(os.getcwd(), topdown=False): if filename in files: return os.path.join(root, filename) raise Exception("File not found") def getWordFreqs(filename): ''' Function that creates a dictionary with the frequency of words in a file. Before the words are counted, each line in the file has any special characters removed with Regex, all letters are set to lowercase, and line breaks are removed. Finally the line is split by word into a list and the words are counted. :param filename: string: the name of the file :return: List: list of WordAndNumber objects. The list is sorted by frequency, decreasing ''' returnList = [] filename = findFile(filename) reader = open(filename, "r", encoding='utf-8') currentLine = "" while True: currentLine = reader.readline() if currentLine == '': break currentLine = re.sub('[^A-Za-z0-9]+', " ", currentLine).lower().replace("\n", "").split(' ') found = False # This loop loops through each word in the 'currentLine' list and checks if a word of that type # has been registered before. If such a word has already been registered, the frequency is incremented. # if not, the word is registered in the returnList as a WordAndNumber object. for word in currentLine: for element in returnList: if element.word == word: element.frequency += 1 found = True break if not found: returnList.append(WordAndNumber(word, 1)) found = False return sorted(returnList, key=lambda element: element.frequency, reverse=True) def getWordsLine(filename, checkWord): ''' Function that creates a list of line numbers where a check word is present The function loops through the file line-by-line. :param filename: string: name of the file to be checked :param checkWord: string: the keyword that is used when checking each line in the file :return: ''' returnList = [] filename = findFile(filename) if not checkWord.startswith(" "): checkWord = " " + checkWord if not checkWord.endswith(" "): checkWord = checkWord + " " reader = open(filename, "r", encoding='utf-8') currentLine = "" lineNumber = 1 while True: currentLine = reader.readline() if currentLine == '': break if checkWord in currentLine: returnList.append(lineNumber) lineNumber += 1 return returnList if __name__ == "__main__": print(getWordsLine("test.txt", "hi")) print(getWordFreqs("11-0.txt"))
[ "thomasbakkenmoe@gmail.com" ]
thomasbakkenmoe@gmail.com
aef286b0f3fd7f3e265e4d48a993b148f77f3f96
ac89c1feea035c5e04c36036f692f059b78f9ce2
/src/device_abstract.py
b0b1cf5497509e0f6ad565cb31626cfc9c4cd21e
[ "Apache-2.0" ]
permissive
urpylka/filesync
3b4a64ad73138d2fdb0b51f394e7310edefafa2b
9167148f27fefdfe56a7b1e2d84479cec56885c5
refs/heads/master
2021-07-11T14:58:53.987749
2021-06-29T06:57:11
2021-06-29T06:57:11
156,345,341
2
3
null
null
null
null
UTF-8
Python
false
false
6,319
py
#! /usr/bin/env python3 # -*- coding: utf-8 -*- # vim:set ts=4 sw=4 et: # Copyright 2018-2019 Artem Smirnov # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # http://www.apache.org/licenses/LICENSE-2.0 # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from threading import Thread from threading import Event class device_abstract(object): """ 1. Как описать общий механизм работы функции, если разные ее исполенения возвращают разный результат? Может возвращать ключ-значение? И это ключ-значение писать в DB? Или функция должна возвращать True/False? 2. А еще наверное нужно передавать какие-то блокировки в downloader/uploader Если передавать большой файл может случится так, что нужно будет притормозить передачу Например, Arming коптера 3. Может сделать фукнции donwload и upload с использованием потоков, таким образом подставив file.open() можно будет писать в файл, а если их направить друг на друга они будут писать без сохранения в local 4. Library for continue a interrupting downloads 5. Загрузка чанками requests https://stackoverflow.com/questions/13909900/progress-of-python-requests-post 6. Надо добавить в вывод "Not implemented method" наименование класса в котором это все вызывается """ @staticmethod def to_string(dic): return "device_abstract://" @staticmethod def get_fields(): return [] def __init__(self, **kwargs): self.kwargs = kwargs self.is_remote_available = Event() t = Thread(target=self._connect, args=()) t.daemon = True t.start() def _connect(self): raise NotImplementedError("Not implemented method '_connect()'") def download(self, device_path, target_stream, chunk_size=1024, offset=0): """ Пока не знаю 1. Функция исполняется в вызывающем потоке 2. Если что-то пошло не так, выбрасывать исключение 3. Если функция возвращает какие-то значения, их нужно передавать по ссылке через аргуемент """ raise NotImplementedError("Not implemented method 'download()'") def upload(self, source_stream, device_path, chunk_size=1024, offset=0): """ Пока не знаю 1. Функция исполняется в вызывающем потоке 2. Если что-то пошло не так, выбрасывать исключение 3. Если функция возвращает какие-то значения, их нужно передавать по ссылке через аргуемент """ raise NotImplementedError("Not implemented method 'upload()'") def get_list(self): """ 1. Функция исполняется в вызывающем потоке 2. Функция должна возвращать список файлов (пустой список, если файлов нет) или, если что-то пошло не так, выбрасывать исключение 3. Функция должна возвращать список словарей {"path":"", "size":"", "hash":""} """ raise NotImplementedError("Not implemented method 'get_list()'") def ls(self, path): """ 1. Функция исполняется в вызывающем потоке 2. Функция должна возвращать список файлов (пустой список, если файлов нет) или, если что-то пошло не так, выбрасывать исключение 3. Функция должна возвращать список путей """ raise NotImplementedError("Not implemented method 'ls()'") def is_dir(self, path): """ """ raise NotImplementedError("Not implemented method 'is_dir()'") def mkdir(self, path): """ """ raise NotImplementedError("Not implemented method 'mkdir()'") def get_size(self, path): """ """ raise NotImplementedError("Not implemented method 'get_size()'") def rename(self, old_path, new_path): """ Вообще одно и тоже, что и move 1. Функция исполняется в вызывающем потоке 2. Если что-то пошло не так, выбрасывать исключение 3. Если функция возвращает какие-то значения, их нужно передавать по ссылке через аргуемент """ raise NotImplementedError("Not implemented method 'rename()'") def delete(self, remote_path): """ 1. Функция исполняется в вызывающем потоке 2. Если что-то пошло не так, выбрасывать исключение 3. Если функция возвращает какие-то значения, их нужно передавать по ссылке через аргуемент """ raise NotImplementedError("Not implemented method 'delete()'")
[ "urpylka@gmail.com" ]
urpylka@gmail.com
d0323d73f123fb6fa6c9c340dcdb2c274040f8c8
42b2b574e400afa875888136bb2a8843c59f2700
/07_Misc/Creative Cheating [crypto] (150)/cheating.py
b6a80a0cc0dd8eb457d0c31b1299b2d30809edba
[]
no_license
yizhimanpadewoniu/CTFLike
d7c572d3c4f39486ff01180c021fde3993a1f252
d822891ede0456d77ea0b191d3726a35d5fec4f5
refs/heads/master
2020-05-22T21:35:22.995073
2018-11-07T12:41:14
2018-11-07T12:41:14
186,530,148
0
1
null
2019-05-14T02:31:41
2019-05-14T02:31:41
null
UTF-8
Python
false
false
1,432
py
import base64 import re import gmpy from Crypto.PublicKey import RSA class RSAPerson(object): def __init__(self, e, p, q): self.n = p * q self.e = e self.p = p self.q = q self.d = long(gmpy.invert(e, (p-1)*(q-1))) self.key = RSA.construct((long(self.n), long(self.e), self.d)) def sign(self, message): return self.key.sign(message, '') def verify(self, message, signature): return self.key.publickey().verify(message, [signature]) def encrypt(self, message): return self.key.publickey().encrypt(message) def decrypt(self, message): return self.key.decrypt(message) alice = RSAPerson( 0x10001, 38456719616722997, 44106885765559411 ) bob = RSAPerson( 0x10001, 49662237675630289, 62515288803124247 ) regex = re.compile(r'SEQ = (\d+); DATA = 0x(.*?)L; SIG = 0x(.*?)L;') packets = [] with open('stream.txt') as lines: for line in lines: decoded = base64.b64decode(line) match = regex.match(decoded).groups() seq = int(match[0]) signature = int(match[2], 16) data = int(match[1], 16) data = bob.decrypt(data) if alice.verify(data, signature): data = chr(data) packets.append(( seq, data, signature )) print ''.join([packet[1] for packet in sorted(packets)])
[ "dengnanyi@163.com" ]
dengnanyi@163.com
ace60bf3e38f65536ff99a19866beecf5e035254
9d51cd52883fc2dec636ad16552592cabd491696
/sentence.py
57c4638e7473c67bc8e8917350fdf80700eda150
[]
no_license
nathan108642/stringtest
6f84ee31458e533232f86334d1e4d75ba4b8dee0
856c6e47f5d3fe44b34fc9691ba9455024792884
refs/heads/master
2020-05-17T20:31:14.919602
2014-02-05T10:26:24
2014-02-05T10:26:24
null
0
0
null
null
null
null
UTF-8
Python
false
false
139
py
start = "hello," question = input ("are you coming out tonight") end = "no,i'm going out to dinner" sentence = start + question = answer
[ "nathanhayter@hotmail.com" ]
nathanhayter@hotmail.com
2e332786982006635b4644f8d511224a688b64c0
ce70410742da246cd1448d2b6ce6e05f8f68a7e2
/manage_cfdi/descarga/migrations/0002_apitbusersrfc.py
1b75e58b41a0bf737938df2ea459be92c2355a22
[]
no_license
agsneutron/balanxa
3255641fa3de3c4be3e00b5fea66c138ad663c64
092c9aa6ec6ce132927e3800663caccfc2db58ce
refs/heads/master
2023-03-30T21:58:17.526923
2021-01-05T05:01:30
2021-01-05T05:01:30
326,866,495
0
2
null
null
null
null
UTF-8
Python
false
false
1,079
py
# Generated by Django 3.1.4 on 2020-12-28 01:58 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('descarga', '0001_initial'), ] operations = [ migrations.CreateModel( name='ApiTbUsersRfc', fields=[ ('user_rfc_id', models.AutoField(primary_key=True, serialize=False, verbose_name='Identificador')), ('user_rfc', models.TextField(blank=True, null=True, verbose_name='RFC')), ('user_rfc_clave', models.TextField(blank=True, null=True, verbose_name='Clave secreta')), ('user_rfc_location_cer', models.TextField(blank=True, null=True, verbose_name='URL del Certificado')), ('user_rfc_location_key', models.TextField(blank=True, null=True, verbose_name='URL de la llave')), ], options={ 'verbose_name': 'RFC de usuario', 'verbose_name_plural': 'RFC de usuarios', 'db_table': 'api_tb_users_rfc', }, ), ]
[ "ags_neutron@hotmail.com" ]
ags_neutron@hotmail.com
7acaed873ed4e401b1f0ffb7abe466b74f9b4a75
be26a7a4b2f6097dc930a3aa696bbdd173512485
/sliding-puzzle/calvins_solution/sliding_puzzle_a_star.py
c95f6a5bace623cdd413f0a323e40be01909cf2c
[]
no_license
jchen8787/inting
c68361e941e01e2fdff8a6a472553cafe7baae04
0ae5caff8e1d33ad22d35ee95c7604066a34acf7
refs/heads/master
2020-05-26T17:52:33.557631
2019-09-10T20:36:08
2019-09-10T20:36:08
188,326,828
0
0
null
null
null
null
UTF-8
Python
false
false
3,952
py
from copy import copy from pdb import set_trace class GameState(object): def __init__(self, board, move_history): self.board = board self.move_history = move_history self._zero_tile_position = None @property def zero_tile_position(self): if self._zero_tile_position: return self._zero_tile_position for i in range(0, len(self.board)): if self.board[i] == 0: self._zero_tile_position = i return self._zero_tile_position @property def available_moves(self): moves = dict() pos = self.zero_tile_position row, col = pos / 3, pos % 3 if row > 0: moves['U'] = 3 * (row - 1) + col if row < 2: moves['D'] = 3 * (row + 1) + col if col > 0: moves['L'] = 3 * row + (col - 1) if col < 2: moves['R'] = 3 * row + (col + 1) return moves @property def cost(self): """The total cost is a sum of the g and h cost, a.k.a the f function """ return self.g_cost + self.h_cost @property def g_cost(self): """The distance away from the source node, a.k.a the root state, which is best represented by the number of steps away from the root state. """ return len(self.move_history) @property def h_cost(self): """Manhattan distance This heuristc function is not the best for sliding puzzle but it is simplest of all. The best heuristic function is Nilsson's sequencing score function. """ final_state = [1, 2, 3, 4, 5, 6, 7, 8, 0] score = 0 for idx, el in enumerate(self.board): score += abs(idx - final_state.index(el)) return score @property def is_solution(self): final_state = [1, 2, 3, 4, 5, 6, 7, 8, 0] for i in range(0, len(self.board)): if final_state[i] != self.board[i]: return False return True def __str__(self): return "".join(map(lambda el: str(el), self.board)) def __hash__(self): return hash(str(self)) def __eq__(self, other): return str(self) == str(other) def find_min_in_queue(queue): """For the lack of a PriorityQueueMap, I am doing an O(n) operation to find the state with the lowest cost """ min_state = None for key in queue: if min_state is None: min_state = queue[key] if min_state.cost > queue[key].cost: min_state = queue[key] return min_state def a_star(board): closed_states = dict() # Initialize the Queue Map open_states = dict() # Initialize root state and put it into the open states root_state = GameState(board, "") open_states[root_state] = root_state while len(open_states) > 0: current_state = find_min_in_queue(open_states) del open_states[current_state] if current_state.is_solution: return current_state closed_states[current_state] = current_state current_pos = current_state.zero_tile_position for dir in current_state.available_moves: new_board = copy(current_state.board) dest = current_state.available_moves[dir] new_board[current_pos], new_board[dest] = new_board[dest], new_board[current_pos] new_move_history = current_state.move_history + dir new_state = GameState(new_board, new_move_history) if closed_states.get(new_state): continue extended_path_cost = new_state.cost if open_states.get(new_state) and open_states.get(new_state).cost <= extended_path_cost: continue open_states[new_state] = new_state return None state = a_star([6, 2, 8, 7, 3, 5, 1, 4, 0]) print state print state.move_history print len(state.move_history)
[ "jusjsc@gmail.com" ]
jusjsc@gmail.com
88f88a537c87284e71ef254d24a05d22fc3a9233
6a928130337dafece1a6158badd00d1d46571003
/reportForm/wsgi.py
28a489cea41932132be6da890e260ca78c6ee72b
[]
no_license
Yanl05/reportForm
bb5a36cff3fac3aca76b5bc50c92fe54282250a8
45a915b29102c1f49035df93217782ea563cdb9f
refs/heads/master
2023-04-18T00:40:19.355040
2021-04-29T14:37:59
2021-04-29T14:37:59
362,485,297
0
0
null
null
null
null
UTF-8
Python
false
false
395
py
""" WSGI config for untitled project. It exposes the WSGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.2/howto/deployment/wsgi/ """ import os from django.core.wsgi import get_wsgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'reportForm.settings') application = get_wsgi_application()
[ "756593069@qq.com" ]
756593069@qq.com
29cd5aa3c4e1875cf4d2d691c2218d861a2d333c
7e4460c85790fae2d470182732289bcd1b8777b2
/Process/process_meshes.py
1ea42ad249869c9afd8713ee9ab0cb63fbd9752a
[]
no_license
khamukkamu/swconquest-msys
5b23654c8dd2e8b2f25bc7914252eedc05a5cc1e
71337a4ae9c507b9440e84cf49d31fc67a781978
refs/heads/master
2021-04-29T19:00:10.389224
2019-05-01T15:11:11
2019-05-01T15:11:11
121,704,753
1
1
null
2018-02-16T01:40:58
2018-02-16T01:40:58
null
UTF-8
Python
false
false
1,015
py
import string from header_common import * from module_info import * from module_meshes import * from process_common import * from process__swyhelper import * def save_meshes(): ofile = open(export_dir + "meshes.txt","w") ofile.write("%d\n"%len(meshes)) for i_mesh in xrange(len(meshes)): mesh = meshes[i_mesh] ofile.write("mesh_%s %d %s %s %s %s %s %s %s %s %s %s\n"%(mesh[0],mesh[1],replace_spaces(mesh[2]),swytrailzro(mesh[3]),swytrailzro(mesh[4]),swytrailzro(mesh[5]),swytrailzro(mesh[6]),swytrailzro(mesh[7]),swytrailzro(mesh[8]),swytrailzro(mesh[9]),swytrailzro(mesh[10]),swytrailzro(mesh[11]))) ofile.close() def save_python_header(): if (wb_compile_switch): ofile = open("./IDs/ID_meshes_wb.py","w") else: ofile = open("./IDs/ID_meshes_mb.py","w") for i_mesh in xrange(len(meshes)): ofile.write("mesh_%s = %d\n"%(meshes[i_mesh][0],i_mesh)) ofile.write("\n\n") ofile.close() print "Exporting meshes..." save_python_header() save_meshes()
[ "swyterzone@gmail.com" ]
swyterzone@gmail.com
1c4a53ccb57b78886be553a6979ec66c482364a8
3805d27bbd55a40594d8d5aebe9aa92f299e5a7f
/data_builder/mhdparser.py
d18b7f1b546ef52f3d1271c5292d11f852136cd9
[]
no_license
novageno/lung
7398b564d266f9401a43daac86e2ddc8ea497170
4ac74f842f07e1a411644e3a8a6d57aee9e0af4e
refs/heads/master
2021-01-16T19:23:43.769888
2017-08-13T06:04:30
2017-08-13T06:04:30
100,157,347
1
0
null
null
null
null
UTF-8
Python
false
false
190
py
# -*- coding: utf-8 -*- """ Created on Sat Aug 12 18:22:44 2017 @author: genonova """ import simpleITK as sitk class MhdParser(): def parse_mhd(self,mhd_dir):
[ "lilhope@163.com" ]
lilhope@163.com
0b1fded14590f90a53b245d8b128a370c4450fac
a4641a95c32923d667e9586f1bf6afef4bfbea51
/finder/forms.py
c864b61b51b32a1007725f096a8aa6392aecad1a
[]
no_license
yumajen/find_news_app
9ac7461e185f17e6a379668987605cad68d55b87
5d7463083df601c5c011c1328f698f91cf27513b
refs/heads/master
2020-06-13T18:48:37.620504
2019-08-12T13:33:50
2019-08-12T13:33:50
194,754,784
0
0
null
2019-08-12T13:26:00
2019-07-01T23:16:27
Python
UTF-8
Python
false
false
1,078
py
from django import forms from .models import NewsSite from .models import Category import datetime class SearchConditions(forms.Form): site_name = forms.ModelChoiceField( label='サイト名:', empty_label='選択してください', queryset=NewsSite.objects.all() ) # TODO: カテゴリは選択したサイトに応じて動的に変化させる category_name = forms.ModelChoiceField( label='カテゴリ:', empty_label='選択してください', queryset=Category.objects.all() ) search_word = forms.CharField( label='検索ワード:', max_length=255 ) search_term_start = forms.DateField( label='取得期間:', widget=forms.DateInput( attrs={ "type": "date" } ), initial=datetime.date.today ) search_term_end = forms.DateField( widget=forms.DateInput( attrs={ "type": "date" } ), initial=datetime.date.today )
[ "noreply@github.com" ]
noreply@github.com
3c3a5c6f1ca843658cc6f58129f94c4f986fe500
83fb1e03b3c83cd6b794f5ac7b3705ed439a25a0
/temperature chart.py
608f03e6f5017da81a58acdf16f8dffc3dc6e8af
[]
no_license
lenatester100/class_assignment
e15d57c3f6288d7a5d459e43ee45e0b1506dd8e3
859340822bbd43318486ebff84e00a047d507fa7
refs/heads/master
2021-01-11T14:24:41.709114
2017-02-17T00:36:53
2017-02-17T00:36:53
81,388,563
0
0
null
null
null
null
UTF-8
Python
false
false
232
py
def main(): print("celsius", "fahrenheit") a=0 while a<=100: a = a + 5 celsiustemp = a fahrenheittemp = (celsiustemp *9/5+32) print(celsiustemp," ", fahrenheittemp) main()
[ "noreply@github.com" ]
noreply@github.com
1a1bd3525f993dcb297270edc0c3315f6f3a3478
0e7b1f88ad2d8d37e1f4071209a41a3c7e7c4172
/pyexec/flask/pytestoutput.py
7be7ea347f45c308459365a643e38d3093cdf420
[]
no_license
veryfreebird/codebase
a74a524b526f2ac4418b0a08485c777d3a6235ff
32f80b56e7b1d9aff4cfef27cd3b75063667b97e
refs/heads/master
2023-08-31T11:40:31.981743
2023-08-28T08:11:36
2023-08-28T08:11:36
53,237,893
3
0
null
null
null
null
UTF-8
Python
false
false
350
py
import pytest from datetime import datetime def test_add(): assert 1 == 2 def gen_report_name(): prefix = '测试报告' ts = datetime.now().strftime('%Y-%m-%d-%H-%M-%S') return prefix + ts + '.html' if __name__ == '__main__': report_name = gen_report_name() pytest.main([f'--html=output/{report_name}'])
[ "79344359@qq.com" ]
79344359@qq.com
a55f91c3b4e428b323ddb4834febff18bff53cb7
ca7aa979e7059467e158830b76673f5b77a0f5a3
/Python_codes/p02818/s319321320.py
ec787c89f517dd3576a0c30e3d24e3bf48cf1b60
[]
no_license
Aasthaengg/IBMdataset
7abb6cbcc4fb03ef5ca68ac64ba460c4a64f8901
f33f1c5c3b16d0ea8d1f5a7d479ad288bb3f48d8
refs/heads/main
2023-04-22T10:22:44.763102
2021-05-13T17:27:22
2021-05-13T17:27:22
367,112,348
0
0
null
null
null
null
UTF-8
Python
false
false
215
py
# ABC149 # B Greesy Takahashi # takはA枚、aokiはB枚、TAKはK回 a, b, k = map(int, input().split()) if k > a: if k - a > b: print(0,0) else: print(0,b - (k - a)) else: print(a-k,b)
[ "66529651+Aastha2104@users.noreply.github.com" ]
66529651+Aastha2104@users.noreply.github.com
145446eb5fc6d444c7b5377f5346c41ba16bce09
e74747ff1dd05f01163e6e903d9364e84b8febf5
/src/main.py
f5d968413dcfc5994a510677aab563a6b79f5001
[ "MIT" ]
permissive
Sdator/bigfootUpData
5b860665336756b6ec1e8f836fa59fa214463a12
b655021d97f35005f9881c9d1841b3d546b12088
refs/heads/main
2023-08-25T12:20:53.723878
2021-09-24T12:57:46
2021-09-24T12:57:46
403,027,825
0
0
null
null
null
null
UTF-8
Python
false
false
8,024
py
# -*- coding: utf-8 -*- ''' by 绝 2019.10.6 QQ 250740270 本程序用于自动更新大脚插件 用到的第三方包 pyinstaller # 打包exe aiohttp # 异步http通信 conda install --name wow aiohttp pyinstaller -y # 安装到wow环境中 生成二进制文件 -F 表示生成单个可执行文件 -w 表示去掉控制台窗口,这在GUI界面时非常有用。不过如果是命令行程序的话那就把这个选项删除吧! -i 表示可执行文件的图标 pyinstaller -F -w 大脚插件自动更新_异步.py -i wow.ico pyinstaller -D 大脚插件自动更新_异步.py # 查错用 导出当前环境 conda env export > conda_vm.yaml 导入环境 conda env create -f conda_vm.yaml ''' import json import sys import os import re # 正则 import zipfile # 解包 import tkinter as tk # 组件 from tkinter import messagebox # 选择框 from tkinter import filedialog # 选择框 import asyncio # 异步 import aiohttp # 引用类型可以直接被函数读取并修改 配置信息 = dict() 配置文件 = "config.json" 预设配置信息 = { "游戏路径": "", "当前版本": "2.5.2.99", "线程": 10, "历史": ['http://wow.bfupdate.178.com/BigFoot/Interface/classic/Interface.1.13.2.18.zip', ] } # 界面交互 class WindowGUI(): __root = tk.Tk() __root.withdraw() # 隐藏Tk窗口 __root.attributes("-topmost", True) # 信息框 @staticmethod def msg(title, msg): # 为了兼容之前的代码 做一个返回值转换 isOK = True if messagebox.askquestion( title, msg) == "yes" else False return isOK @staticmethod def exit(title="错误", msg="退出程序",): messagebox.showinfo(title, msg) sys.exit() @staticmethod def 选择框(title, **args): 选择的文件夹 = filedialog.askdirectory( title=title, **args) if not 选择的文件夹: defpath = os.getcwd() if WindowGUI.msg("提示", f"没有选择目录是否才用当前目录?\n当前目录为:{defpath}"): return defpath else: WindowGUI.exit("再见", "等想好了目录再来找我吧!") if sys.path[0] == 选择的文件夹: if not WindowGUI.msg("提示", f"检测到选择目录和当前目录相同,是否要下载到当前目录?\n选择目录为:{选择的文件夹}"): WindowGUI.exit("再见", "等想好了目录再来找我吧!") return 选择的文件夹 def 组合地址(版本号): return "http://wow.bfupdate.178.com/BigFoot/Interface/classic/Interface.%s.zip" % (版本号) def 选择游戏目录(): # 判断key是否存在的正规写法 # 如果配置中不存在路径则触发路径选择 if not('游戏路径' in 配置信息) or not(os.path.exists(os.path.dirname(配置信息["游戏路径"]))): '''打开选择文件夹对话框''' 选择的文件夹 = WindowGUI.选择框(r'选择魔兽世界根目录如:X: \Games\World of Warcraft') # 获得选择的文件夹 配置信息["游戏路径"] = os.path.normcase(选择的文件夹 + "\\_classic_") def 读入配置(path): # 路径合法性 配置文件 = os.path.normcase(path) # 文件不存在采用预设配置 if not os.path.isfile(配置文件): print("文件不存在返回预设配置") return 预设配置信息 # 读入配置 # 以utf8打开文件 并转为json with open(配置文件, "r+", encoding='utf-8') as json_file: # 检测json的合法性 try: config = json.load(json_file) # 如果是字典类型才是正确的 if type(config) == dict: return config print("json解析错误 使用预设配置") return 预设配置信息 except json.decoder.JSONDecodeError as e: print("json解析失败 使用预设配置,错误信息:", e) return 预设配置信息 def 写出配置(data): # 覆盖原有 with open(配置文件, "w", encoding='utf-8') as json_file: # 把dict对象转为json并允许非ASCII字符 json_file.write(json.dumps(data, ensure_ascii=False)) async def fetch(session, url): async with session.head(url) as resp: if resp.status == 200: return url async def 获取最新版本(): # 版本号 分割 并转为整数 a, b, c, d = [int(i) for i in 配置信息["当前版本"].split(".")] # 使用会话 async with aiohttp.ClientSession() as session: # 创建异步任务列表 tasks = [fetch(session, 组合地址(f'{a}.{b}.{c}.{d+i}')) for i in range(配置信息["线程"])] # 异步访问 urls = await asyncio.gather(*tasks) # 去除 None 结果 历史 = [url for url in urls if url] if not len(历史): print("找不到可用的版本,尝试加大线程数量或直接修改配置“当前版本”为最近的一个版本的近似数") WindowGUI.msg( u"错误", u"找不到可用的版本,尝试加大线程数量或直接修改配置“当前版本”为最近的一个版本的近似数。") return # 正则匹配出版本号 最新版本 = re.match(r'.*(\d+\.\d+\.\d+\.\d+)', 历史[-1]).group(1) # 历史版本合并 去重复 排列 配置信息["历史"] = sorted(set(配置信息["历史"] + 历史)) if 配置信息["当前版本"] == 最新版本: WindowGUI.msg(u"提示", u"当前已是最新版本,无需更新") return return 历史[-1] async def 下载插件(url): async with aiohttp.ClientSession() as session: async with session.get(url) as resp: 文件大小 = int(resp.headers['Content-Length']) 保存路径 = os.path.normcase("%s.zip" % (配置信息["当前版本"])) with open(保存路径, 'wb') as fd: while True: # 读取文件流 chunk = await resp.content.read(文件大小) if not chunk: return 保存路径 fd.write(chunk) def 打开文件夹(paht): if os.path.exists(paht): os.system("start " + paht) else: WindowGUI.exit("找不到目录", "压缩包存在问题或解压失败") # 解压程序 def 解压(file, path): try: # 打开压缩包 z = zipfile.ZipFile(file, "r") # 解压到指定位置 z.extractall(path) z.close() isOpen = WindowGUI.msg(u"提示", u"安装完成! 是否打开文件夹?") if isOpen: 打开文件夹(path) except FileNotFoundError as e: print("解压失败找不到文件:", e) WindowGUI.exit(u"错误", u"找不到压缩文件,检测路径或文件名是否正确") except zipfile.BadZipFile as e: print("文件格式错误:", e) WindowGUI.exit(u"错误", u"文件格式错误,检测是否正确的zip文件") async def main(): # 更新全局变量的值 由于全局变量无法直接赋值 但可以执行其方法 引用类型的元素可以赋值 配置信息.update(读入配置(配置文件)) 选择游戏目录() url = await 获取最新版本() if url: name = await 下载插件(url) 解压(name, 配置信息["游戏路径"]) 写出配置(配置信息) if __name__ == '__main__': # 错误处理 try: asyncio.run(main()) except aiohttp.client_exceptions.ClientConnectorError as e: print("远程计算机拒绝网络连接:", e) except aiohttp.client_exceptions.InvalidURL as e: print("地址格式格式有误", e) except AssertionError as e: print("网络状态非200", e)
[ "250740270@qq.com" ]
250740270@qq.com
d846750416962cc14802ebec76b38e76d668611c
b3eef5d20818cbb2462acaedac3544a703987a60
/maze_generator/render.py
64e4f7d2262508990b92d4380b483585956fe865
[]
no_license
JohnnyDeuss/maze-generator
5a36cf69e7d51c8e09a40f77d36fa13054d13e95
ae7107d787e80206d96d2a92776f606ac3ce5210
refs/heads/main
2023-02-15T01:38:54.508688
2021-01-02T22:06:49
2021-01-02T22:08:36
326,081,260
2
0
null
null
null
null
UTF-8
Python
false
false
1,381
py
from PIL import Image, ImageDraw def render_maze(g, path_width=20, wall_width=2, border_width=4, output_file="maze.png"): """ Render a maze graph and output it to the given file. """ w, h = list(g.nodes)[-1] w += 1 h += 1 img_w = w * path_width + (w - 1) * wall_width + 2 * border_width img_h = h * path_width + (h - 1) * wall_width + 2 * border_width img = Image.new("1", (img_w, img_h)) draw = ImageDraw.Draw(img) # Draw entrance draw.rectangle( [0, border_width, border_width, border_width + path_width - 1], fill=1 ) # Draw exit draw.rectangle( [ img_w - border_width, img_h - path_width - border_width, img_w, img_h - border_width - 1, ], fill=1, ) def draw_line(a, b): # Ensure `a` is the upper left point. if b[0] < a[0] or b[1] < a[1]: a, b = b, a x_a, y_a = a x_b, y_b = b x_a = border_width + x_a * (path_width + wall_width) y_a = border_width + y_a * (path_width + wall_width) x_b = border_width + x_b * (path_width + wall_width) + path_width - 1 y_b = border_width + y_b * (path_width + wall_width) + path_width - 1 draw.rectangle([x_a, y_a, x_b, y_b], fill=1, outline=1) for e in g.edges: draw_line(*e) return img
[ "johnnydeuss@gmail.com" ]
johnnydeuss@gmail.com
d3e98871213596ac415e24442f42d914dfd5ad66
28c06b1ebc6b361b96be62f5274c3f6affa991ab
/sat_modules/utils.py
daffe19f47096fd638002e1edfa21655ac17d88a
[]
no_license
caifti/sat
c5df52fa86c2de960f11ab267f38c4ebc38ca489
54380d66b75d46c37e48269b25682cce50157f5f
refs/heads/master
2020-09-09T09:14:30.010315
2019-06-21T07:50:22
2019-06-21T07:50:22
221,409,359
0
0
null
2019-11-13T08:27:24
2019-11-13T08:27:24
null
UTF-8
Python
false
false
3,492
py
# -*- coding: utf-8 -*- # Copyright 2018 Spanish National Research Council (CSIC) # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """ Satellite utils Author: Daniel Garcia Diaz Date: May 2018 """ #Submodules from sat_modules import config #APIs import zipfile, tarfile import argparse import numpy as np import os, shutil import json import datetime import utm from netCDF4 import Dataset from six import string_types def valid_date(sd, ed): """ check if the format date input is string("%Y-%m-%d") or datetime.date and return it as format datetime.strptime("YYYY-MM-dd", "%Y-%m-%d") Parameters ---------- sd(start_date) : str "%Y-%m-%d" ed(end_date) : str "%Y-%m-%d" Returns ------- sd : datetime datetime.strptime("YYYY-MM-dd", "%Y-%m-%d") ed : datetime datetime.strptime("YYYY-MM-dd", "%Y-%m-%d") Raises ------ FormatError Unsupported format date ValueError Unsupported date value """ if isinstance(sd, datetime.date) and isinstance(ed, datetime.date): return sd, ed elif isinstance(sd, string_types) and isinstance(ed, string_types): try: sd = datetime.datetime.strptime(sd, "%Y-%m-%d") ed = datetime.datetime.strptime(ed, "%Y-%m-%d") if sd < ed: return sd, ed else: msg = "Unsupported date value: '{} or {}'.".format(sd, ed) raise argparse.ArgumentTypeError(msg) except: msg = "Unsupported format date: '{} or {}'.".format(sd, ed) raise argparse.ArgumentTypeError(msg) else: msg = "Unsupported format date: '{} or {}'.".format(sd, ed) raise argparse.ArgumentTypeError(msg) def valid_region(r): """ check if the regions exits Parameters ---------- r(region) : str e.g: "CdP" Raises ------ FormatError Not a valid region """ if r in config.regions: pass else: msg = "Not a valid region: '{0}'.".format(r) raise argparse.ArgumentTypeError(msg) def path(): """ Configure the tree of datasets path. Create the folder and the downloaded_files file. Parameters ---------- path : datasets path from config file """ file = 'downloaded_files.json' list_region = config.regions local_path = config.local_path try: with open(os.path.join(local_path, file)) as data_file: json.load(data_file) except: if not (os.path.isdir(local_path)): os.mkdir(local_path) dictionary = {"Sentinel-2": {}, "Landsat 8": {}} for region in list_region: os.mkdir(os.path.join(local_path, region)) dictionary['Sentinel-2'][region] = [] dictionary['Landsat 8'][region] = [] with open(os.path.join(local_path, 'downloaded_files.json'), 'w') as outfile: json.dump(dictionary, outfile)
[ "garciad@ifca.unican.es" ]
garciad@ifca.unican.es
3f21ba69b37a994780c6046ac37ba58a78d26472
29e4005d970b66b1640e9570590addca66d8a53a
/czblog/settings.py
35cdb6bd06d2d9225e8a6514798180faa3040b6c
[]
no_license
cz9025/czblog
ec07582fdd06d77349fd518b2a7d94daeffc535b
50b712607bbe33c687e790dcc93a82e1f32f58bd
refs/heads/master
2020-04-13T04:25:43.642086
2020-01-03T09:38:04
2020-01-03T09:38:04
162,960,443
0
0
null
null
null
null
UTF-8
Python
false
false
3,955
py
# -*- coding: utf-8 -*- """ Django settings for czblog project. Generated by 'django-admin startproject' using Django 1.9. For more information on this file, see https://docs.djangoproject.com/en/1.9/topics/settings/ For the full list of settings and their values, see https://docs.djangoproject.com/en/1.9/ref/settings/ """ import os # Build paths inside the project like this: os.path.join(BASE_DIR, ...) BASE_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) # Quick-start development settings - unsuitable for production # See https://docs.djangoproject.com/en/1.9/howto/deployment/checklist/ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'cfam^kgsdky$rs$is#s(5zklw1*7%b75#oslzjshlzz2$m@79z' # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True # DEBUG = False ALLOWED_HOSTS = ['*'] # Application definition INSTALLED_APPS = [ 'django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'xadmin', 'crispy_forms', 'blog', 'center', 'news', 'shop', 'resume', 'myblog', 'interface', 'csdn', ] MIDDLEWARE_CLASSES = [ 'django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', # 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.auth.middleware.SessionAuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', ] ROOT_URLCONF = 'czblog.urls' TEMPLATES = [ { 'BACKEND': 'django.template.backends.django.DjangoTemplates', 'DIRS': [os.path.join(BASE_DIR, "templates")], 'APP_DIRS': True, 'OPTIONS': { 'context_processors': [ 'django.template.context_processors.debug', 'django.template.context_processors.request', 'django.contrib.auth.context_processors.auth', 'django.contrib.messages.context_processors.messages', 'django.template.context_processors.media', ############wenjian ], }, }, ] WSGI_APPLICATION = 'czblog.wsgi.application' # Database # https://docs.djangoproject.com/en/1.9/ref/settings/#databases DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'czblog', 'USER': 'root', 'PASSWORD': '123456', 'HOST': '127.0.0.1', 'PORT': '3306', } } # Password validation # https://docs.djangoproject.com/en/1.9/ref/settings/#auth-password-validators AUTH_PASSWORD_VALIDATORS = [ { 'NAME': 'django.contrib.auth.password_validation.UserAttributeSimilarityValidator', }, { 'NAME': 'django.contrib.auth.password_validation.MinimumLengthValidator', }, { 'NAME': 'django.contrib.auth.password_validation.CommonPasswordValidator', }, { 'NAME': 'django.contrib.auth.password_validation.NumericPasswordValidator', }, ] # Internationalization # https://docs.djangoproject.com/en/1.9/topics/i18n/ # LANGUAGE_CODE = 'zh-hans' LANGUAGE_CODE = 'en-us' TIME_ZONE = 'Asia/Shanghai' USE_I18N = True USE_L10N = True USE_TZ = False # Static files (CSS, JavaScript, Images) # https://docs.djangoproject.com/en/1.9/howto/static-files/ AUTH_USER_MODEL = 'center.UserInfo' STATIC_URL = '/static/' STATICFILES_DIRS = [ os.path.join(BASE_DIR, 'static'), ] # 放服务器上时,需打开 已作废 # STATIC_ROOT='/usr/local/src/webroot/blog/static/' # MEDIA_ROOT = '/usr/local/src/webroot/media/' # windows上需关闭 # STATIC_ROOT=os.path.join(BASE_DIR,'static') MEDIA_ROOT=os.path.join(BASE_DIR,'media') MEDIA_URL = '/media/'
[ "897308902@qq.com" ]
897308902@qq.com
36d479e9d62294ff7f9e197330fafe40334c2523
b5e5182e0031b84af7eeccd8cdbf9dac457a4c03
/plugins/InformaticaPlugin/operators/execute_profile.py
96f5e257636bff5b44e96145725ba289b4b4872b
[ "MIT" ]
permissive
consag/informatica-airflow-plugin
0c8f86dd4b466a74515f65b648489dd95b1e75ef
50c2b713d71b80eb0a5202c84b9e1d3173f0512f
refs/heads/master
2023-05-12T09:10:56.966475
2022-09-16T21:58:02
2022-09-16T21:58:02
219,062,936
3
0
MIT
2023-05-08T20:32:30
2019-11-01T21:08:28
Python
UTF-8
Python
false
false
1,551
py
from airflow.models import BaseOperator from airflow import utils as airflow_utils, AirflowException from execution import runProfile from InformaticaPlugin.operators import available_arguments import os class ExecuteProfile(BaseOperator): @airflow_utils.apply_defaults def __init__(self, profile_path, **kwargs): self.infa_arguments = [] self.pre_command = None for key, value in kwargs.items(): if key == 'target': self.pre_command = '. ' + os.environ.get('configDir', '.') + '/scheduler_env.' + value + '.sh' else: if key in available_arguments: self.infa_arguments.append(available_arguments[key] + " " + value) super(ExecuteProfile, self).__init__( **kwargs) def execute(self, context): print("dag: " + self.dag.full_filepath) print("dag_id: " + self.dag_id) print("task_type: " + self.task_type) print("task id: " + self.task_id) print("infa_arguments: " + ' '.join(self.infa_arguments)) if self.pre_command is None: print("no pre_command provided.") else: print("pre_command: " + self.pre_command) infa = runProfile.ExecuteInformaticaProfile(self.infa_arguments, log_on_console=False, pre_command=self.pre_command) result = infa.runit(infa.arguments) if result.rc != 0: raise AirflowException("RunProfile failed: " + result.message)
[ "jac@jacbeekers.nl" ]
jac@jacbeekers.nl
0d7d6b9040e1904e68f4c40434dae2fa937d9729
61a4a77fb2347172390a8528a068207d50804d61
/python/function.py
9f4ddc9d868a1a66f2efc2163aba671fecd73d89
[]
no_license
king636/myLearning
84180c05dbc418bd30919928ed95fb3788d191d8
7a3c3eee70bbe425c3575b2ac4837d9e4dabed1a
refs/heads/master
2021-09-06T04:25:27.255477
2018-02-02T09:41:54
2018-02-02T09:41:54
113,003,979
2
0
null
null
null
null
UTF-8
Python
false
false
11,305
py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- # 函数用法 # python内置的函数 # 1. abs()取绝对值,js也有内置函数:Math.abs() # 20 print(abs(-20)) # 参数不对,报错:TypeError: abs() takes exactly one argument (2 given) # js不报错 # abs(-20,1) # 参数类型不对,报错:TypeError: bad operand type for abs(): 'str' # js的Math.abs()参数类型不对,返回NaN # abs('a') # 内置函数2:max,取最大值 js用Math.max() # 8 print(max(1,8,3,5)) # 内置函数3: # int() # str() # bool() # 1 print(int('1')) # 报错:ValueError: invalid literal for int() with base 10: 'a' # print(int('a')) # 1234 print(str(1234)) # None print(str(None)) # True print(bool(1)) # False print(bool(None)) # False print(bool('')) # True print(bool('a')) # 内置函数:hex() # 0xa print(hex(10)) ############################################ # 定义函数 # 函数不写return时,返回None.而js是undefined def my_abs(x): if x >= 0: return x; else: return -x; # 99 print(my_abs(-99)) ## 函数如何在别的文件中导入?参看function_my_abs.py,使用from function import my_abs来导入 # js中我只知道只能通过宿主(比如浏览器),在html文件中导入js文件,然后使用其函数. # 单纯的js文件应该无法像python一样导入其他文件的函数,所以python更像c,可以导入头文件.js是脚本,寄托于宿主. # 空函数与pass关键字,占位用,确保代码正常运行 def fun(): pass # pass放到判断中,代表什么也不做.如果不用pass,会有语法错误 x = 20 if x > 10: pass # my_abs函数扩展,增加类型检查,以及抛出异常 # 检查参数是否是int或者float类型,如果不是,抛出TypeError异常 oprand:操作数 # js中检查x的类型:typeof x !== 'number' # python使用raise Error对象来抛出异常,js使用throw('...') def my_abs_extend(x): if not isinstance(x,(int,float)): raise TypeError('bad operand type') if x >= 0: return x else: return -x #TypeError: bad operand type # my_abs_extend('a') ### 返回多个值,事实上与js中的解构赋值类似,分析如下: import math # move函数根据点移动时的坐标 位移 和角度,算出新的点的坐标 # 这里参数angle=0定义了默认值,默认参数的使用:与js类似,必须写在后面 def move(x, y, step, angle=0): nx = x + step * math.cos(angle) ny = y - step * math.sin(angle) return nx, ny x, y = move(100, 100, 60, math.pi / 6) # 151.96152422706632 70.0 print(x,y) # 事实上相当于以下,move函数返回的是tuple r = move(100, 100, 60, math.pi / 6) # (151.96152422706632, 70.0) print(r) # 这里的作用就是解构赋值,js中数组对应的解构赋值:[a,b] = [151.96152422706632,70.0] (a,b) = r # 151.96152422706632 70.0 print(a,b) # 与js一样(不过js是用[]),解构赋值的层级也需要对应 (a,b,(c,d)) = (1,2,(3,4)) # c = 3 print('c = ' + str(c)) # 当然也可以和js一样,python的list也可以解构赋值(tuple和list基本是一样的,除了不可变性质) [a,b,[c,d]] = [10,20,[30,40]] print('c = ' + str(c)) # js可以从对象中解构赋值,python是否可以从dict解构赋值呢? # 报错:SyntaxError: can't assign to literal # {a,b} = { # 'name':'nick', # 'age':30, # } # 根据上面报错的提示,不能赋值给literal, 修改为tuple (a,b) = { 'name':'nick', 'age':30, } # 赋值的是key:name: name,age: age print('name: ' + a + ',age: ' + b) # 不像js,python的dict只能解构到key,要获取value,如下: person = { 'name':'nick', 'age':30, } (a,b) = person # name: nick print('name: ' + person[a]) # 等价于 # name: nick print('name: ' + person['name']) # 加入层次和使用list person = { 'name':'nick', 'age':30, 'address':{ 'city':'Xiamen', 'country':'China' } } # 报错:ValueError: too many values to unpack (expected 2) # 所以dict的解构不能使用层次,那么就需要二次解构了 # (a,b,(c,d)) = person (a,b,c) = person # address print(c) (a,b) = person[c] # city # 这里经过二次解构 print(a) ### 上面的tuple换成list也是一样的,说明list和tuple在解构赋值时是一样用法 [a,b,c] = person # address print(c) [a,b] = person[c] # city print(a) ### 默认参数,可变参数,关键字参数,命名关键字参数 ## 注意默认参数可能会发生改变的情况 def abc(L = []): L.append('END') return L # 第一次调用默认参数:['END'] print(abc()) # 第二次调用:['END', 'END'] # L每次调用后被改变,因为函数定义的时候python就算出默认参数变量L的值. # js也有类似的问题,所以默认参数一定要指向不可变对象 print(abc()) # 默认参数的问题可以修改如下: def xyz(L = None): if L is None: L = [] L.append('END') return L # ['END'] print(xyz()) # ['END'] print(xyz()) ## 可变参数 # 定义一个函数,参数可传入list或者tuple # 通过导入collections模块,判断参数是否是Iterable(list或tuple), 如果是才能使用for...in import collections def calc(numbers): if not isinstance(numbers,collections.Iterable): raise TypeError('is not iterable') sum = 0 for n in numbers: sum = sum + n * n return sum # 用法: # 14 print(calc([1,2,3])) # 14 print(calc((1,2,3))) # TypeError: is not iterable # print(calc(1)) # 改为可变参数,加* def calc(*numbers): if not isinstance(numbers,collections.Iterable): raise TypeError('is not iterable') sum = 0 for n in numbers: sum = sum + n * n return sum # 14 print(calc(1,2,3)) # 不能再传入list了,报错:TypeError: can't multiply sequence by non-int of type 'list' # print(calc([1,2,3])) # 这时要改为: li = [1,2,3] # 14 print(calc(li[0],li[1],li[2])) # 再改:*list 表示把list的所有元素作为可变参数传入 # 14 print(calc(*li)) # 0 print(calc()) ## 关键字参数 ** # 关键字参数组装成dict,参数个数任意 def registe(name,age,**other): print('name:',name,',age:',',other:',other) # name: nick ,age: ,other: {} registe('nick',30) # name: nick ,age: ,other: {'city': 'Xiamen', 'country': 'China'} registe('nick',30,city='Xiamen',country='China') # 关键字参数用处:比如注册时,输入必须信息后,还可以输入额外信息 # 与可变参数类型: extra = { 'city':'Xiamen', 'country':'China' } # name: Bob ,age: ,other: {'city': 'Xiamen', 'country': 'China'} registe('Bob',50,city=extra['city'],contry=extra['country']) # 简化用法 # name: Bob ,age: ,other: {'city': 'Xiamen', 'country': 'China'} registe('Bob',50,**extra) ## 命名关键字参数 # 对于关键字参数,如果要限制参数名字,使用命名关键字参数,用*隔开 def registe(name,age,*,city,country): print(name,age,city,country) # TypeError: registe() missing 2 required keyword-only arguments: 'city' and 'country' # 可见,命名关键字参数个数一致,且参数名要写,还要写对 # registe('Nick',30) # nick 20 Shanghai China registe('nick',20,city='Shanghai',country='China') # 如果用可变参数,那么就不用*隔开了 def registe(name,age,*other,city,country): print(name,age,other,city,country) # cathy 18 () Xian China registe('cathy',18,city='Xian',country='China') # cathy 18 ((1, 2, 3),) Xian China registe('cathy',18,(1,2,3),city='Xian',country='China') # kate 20 () Xiamen China registe('kate',20,**extra) #### 参数组合,上面的参数可以组合使用,但是必须确保顺序: # 必选参数 默认参数 可变参数 命名关键字参数 关键字参数 # 注意:命名关键字参数在关键字参数前面 def f1(a,b,c=0,*change,d,**extra): print('a=',a,'b=',b,'c=',c,'change=',change,'d=',d,'extra=',extra) def f2(a,b,*,d,**extra): print('a=',a,'b=',b,'d=',d,'extra=',extra) # a= 1 b= 2 c= 3 change= ((3, 4),) d= abc extra= {'name': 'nick', 'age': 30} f1(1,2,3,(3,4),d='abc',name='nick',age=30) # 命名关键字d是不能省略的 # TypeError: f1() missing 1 required keyword-only argument: 'd' # f1(1,2) # a= 1 b= 2 c= 0 change= () d= 30 extra= {} f1(1,2,d=30) ### 重点来了:任意函数参数都可以通过fun(*arg1,**arg2)来调用,不管实际的参数怎么定义 t = (1,2,3,4) d = {'d':50,'city':'Beijing','country':'China'} # a= 1 b= 2 c= 3 change= (4,) d= 50 extra= {'city': 'Beijing', 'country': 'China'} f1(*t,**d) # 因为有命名关键字参数d,所以dict d中必须有'd' d = {'city':'Beijing','country':'China'} # 报错:TypeError: f1() missing 1 required keyword-only argument: 'd' # f1(*t,**d) ## 递归的使用 # 使用递归要注意防止栈溢出:函数调用是通过栈的数据结构实现的 # 每当进入一个函数调用,栈就会加一层栈帧,当函数退出,栈就减一层栈帧. # 而栈的大小有限,如果递归层级太多(调用函数自身太多次),可能会栈溢出 # 比如使用递归计算阶乘:1*2*3*4*5...*n! def fact(n): if n == 1: return 1 return fact(n - 1) * n # 120 print(fact(5)) # RecursionError: maximum recursion depth exceeded in comparison # print(fact(1000)) ## 防止栈溢出,可以通过尾递归优化 # 尾递归:函数返回的时候调用了函数自身(递归),return的不能是表达式,比如上面的:fact(n - 1) * n是表达式. # 尾递归优化:当满足尾递归时,编译器会对其优化,使得不管递归调用多少次,都只会加一层栈帧 # 上面的递归函数进行改写: def fact_iter(num,product=1): if num == 1: return product return fact_iter(num - 1, num * product) # 120 print(fact_iter(5)) # RecursionError: maximum recursion depth exceeded in comparison # 这里仍然报错,是因为使用的python解释器并没有对尾递归进行优化 # print(fact_iter(1000)) #### 汉诺塔的移动:http://www.baike.com/wiki/%E6%B1%89%E8%AF%BA%E5%A1%94 # 三根柱子A,B,C A上从下往上按从大到小的顺序堆了N个盘子,计算将盘子堆到另外一根柱子上(这里从A到C)所需次数,一次只能移动一个盘子 # 并且大盘子不能堆在小盘子上. # 按照数学归纳法:1个盘子->1次 2个盘子->3次 3个盘子->7次 4个盘子->15次...归纳出:2^n - 1次 # 找规律分析: # 1个盘子:A-->C # 2个盘子:A-->B A-->C B-->C # 3个盘子:A-->C A-->B C-->B A-->C B-->A B-->C A-->C # ... # 过程理解:n个盘子,先从A移动n-1个到B(当前过程是复杂的);A还剩下一个时,从A移动到C;最后从B将n-1个移动到C # 用递归实现: def move(n,a,b,c): if n==1: print('move',a,'-->',c) else: move(n-1,a,c,b) move(1,a,b,c) move(n-1,b,a,c) # move A --> C move(1,'A','B','C') # move A --> B # move A --> C # move B --> C move(2,'A','B','C') # move A --> C # move A --> B # move C --> B # move A --> C # move B --> A # move B --> C # move A --> C move(3,'A','B','C') ### 递归的理解不要陷入死胡同,总结出规律,如果确认是递归问题,那么理清过程去调用函数即可,别陷太深....
[ "ChenBin@yaxon.com" ]
ChenBin@yaxon.com
53dc368768d6cfff5f959b78f918c8c2190e3a95
c8abe556d3f01071d2df5d784746e0ea8a590544
/code/box_plots.py
7d7e5c5bfb3355d5d036e8188730a156d2f1866c
[]
no_license
02450-Intro-to-ML-ETHOTOS/automobiles
0fc0b8816edd074f74b3e75bb56e9908149ecccb
3a9f62e4ea2718188cd5067258be568550a0bed4
refs/heads/master
2023-04-09T19:53:20.335351
2021-04-20T12:30:14
2021-04-20T12:30:14
343,370,253
0
0
null
null
null
null
UTF-8
Python
false
false
1,593
py
from pca import * from scipy.stats import zscore #Extract Attributes wheel-base, length, width, height #dropped_data = dropped_data.iloc [:, [0,1,2,3]] #attributeNames = np.array(dropped_data.columns) X = np.array(dropped_data) #The data matrix should be standardized to have zero mean and #unit standard deviation X = zscore(X, ddof=1) #------------------------------------------------------------------------------ ### BoX of selected colums figure(figsize=(10,7)) boxplot(X) r = np.arange(1,X.shape[1]+1) xticks(r,attributeNames) xticks(rotation = 45) ylabel('m') title('Automobiles - boxplot') show() #------------------------------------------------------------------------------ price_plot = dropped_data.iloc[:, [12]] attributeName = np.array(price_plot.columns) price_plot = np.array(price_plot) ### Boxplot of prices boxplot(price_plot) r = np.arange(1,price_plot.shape[1]+1) xticks(r,attributeName) xticks(rotation = 45) ylabel('dollar') title('Automobiles - boxplot') show() #------------------------------------------------------------------------------ figure(figsize=(25,7)) for c in range(C): subplot(1,C,c+1) class_mask = (y==c) # binary mask to extract elements of class c # or: class_mask = nonzero(y==c)[0].tolist()[0] # indices of class c boxplot(X[class_mask,:]) #title('Class: {0}'.format(classNames[c])) title('Class: '+classNames[c]) xticks(range(1,len(attributeNames)+1), [a[:7] for a in attributeNames], rotation=45) y_up = X.max()+(X.max()-X.min())*0.1; y_down = X.min()-(X.max()-X.min())*0.1 ylim(y_down, y_up) show()
[ "70723194+Erikinol@users.noreply.github.com" ]
70723194+Erikinol@users.noreply.github.com
3a7b75886db5b91af55617c1f0a203df4479f839
fcad3e866d3e3815ba5e7d2c9b5d1ab2bb3917b9
/face.py
258946bff9c751a23f86c487b4fc7b8e497febce
[]
no_license
ranjeetbidwe/Kaggle-Autism
520b20cf55b79bcb29aee28f711bbb46ff9891c7
bda23282cf004cbe48fad86890d66da306ede835
refs/heads/master
2022-04-26T04:30:00.698400
2020-04-27T20:46:21
2020-04-27T20:46:21
null
0
0
null
null
null
null
UTF-8
Python
false
false
7,609
py
import os import glob import time import keras from PIL import Image from os import listdir from shutil import copyfile from os.path import isfile, join from matplotlib import pyplot as plt from keras_vggface.vggface import VGGFace from keras.engine import Input from keras import applications from keras.models import Model import tensorflow as tf import numpy as np from keras.layers import ZeroPadding2D, Convolution2D, MaxPooling2D, Flatten, Dense, Dropout from keras_vggface.utils import preprocess_input from sklearn.metrics import classification_report, confusion_matrix from MLEXPS.MLEXPS import * from keras import backend as K import random random.seed(42) tf.random.set_seed(42) # Providing more training examples within certain distributions of age, gender, and race will increase the model's accuracy. Height = 224 Width = 224 BatchSize = 24 lr_rate=.0015 Version = 5 load_model = False model_path = '' accuracy = 0 accuracyCount = 0 trainableCount = 30 def SaveModelImage(Model, Title): keras.utils.vis_utils.plot_model(Model, to_file=Title, show_shapes=True, show_layer_names=True) return def Summary(Model): print(Model.summary()) return def resnet(): BaseModel = applications.resnet50.ResNet50(weights= None, include_top=False, input_shape= (224,224,3)) last_layer = BaseModel.get_layer('activation_49').output print('here') return model def MakeModel(dlsize): BaseModel = VGGFace(model='senet50', include_top=False, input_shape=(Height, Width, 3), pooling='avg') last_layer = BaseModel.get_layer('avg_pool').output x = keras.layers.Flatten(name='flatten')(last_layer) x = keras.layers.Dense(128, kernel_regularizer = keras.regularizers.l2(l = 0.015), activation='relu')(x) x = keras.layers.Dropout(rate=.4, seed=42)(x) out = keras.layers.Dense(2, activation='softmax', name='classifier')(x) DerivedModel = keras.Model(BaseModel.input, out) # # Everything is trainingable # # Weights are used at init # for layer in DerivedModel.layers: # layer.trainable = True # # # # Everything in the base model is frozen # # Only top layers are trainable # for layer in BaseModel.layers: # layer.trainable = False # base for layer in DerivedModel.layers: layer.trainable = False for layer in DerivedModel.layers[-trainableCount:]: layer.trainable = True DerivedModel.compile(keras.optimizers.Adam(lr=lr_rate), loss='categorical_crossentropy', metrics=['accuracy']) return DerivedModel def clearWeights(model): weights = model.get_weights() for weight in weights: weight = K.zeros(weight.shape, dtype=np.float64) model.set_weights(weights) return model def preprocess_input_new(x): img = preprocess_input(keras.preprocessing.image.img_to_array(x), version = 2) return keras.preprocessing.image.array_to_img(img) class EarlyStoppingAtMinLoss(tf.keras.callbacks.Callback): def __init__(self, trainableCount=30): print('working') super(EarlyStoppingAtMinLoss, self).__init__() self.epochCount = [] self.trainableCount = trainableCount self.max = 0 def on_train_begin(self, logs=None): self.accuracyCount = 0 self.accuracy = 0 def on_epoch_end(self, epoch, logs=None): self.max = len(self.model.layers) print("Ending Epoch") if logs['val_accuracy'] > self.accuracy: self.accuracy = logs['val_accuracy'] self.accuracyCount = 0 else: self.accuracyCount+=1 if self.accuracyCount >= 10 * (len(self.epochCount)+1): self.epochCount.append(epoch) print('Adding train layers') self.accuracyCount = 0 self.trainableCount += 10 if self.trainableCount >= self.max: self.trainableCount = self.max for layer in self.model.layers: layer.trainable = False for layer in self.model.layers[-self.trainableCount:]: layer.trainable = True self.model.compile(keras.optimizers.Adam(lr=lr_rate), loss='categorical_crossentropy', metrics=['accuracy']) print(self.epochCount) if __name__ == "__main__": timestr = time.strftime("%Y%m%d-%H%M%S") model = MakeModel(1024) # model = resnet() # model = clearWeights(model) model.compile(keras.optimizers.Adam(lr=lr_rate), loss='categorical_crossentropy', metrics=['accuracy']) model.summary() TrainPath = 'D:/Autism-Data/Kaggle/v' + str(Version) + '/train' ValidPath = 'D:/Autism-Data/Kaggle/v' + str(Version) + '/valid' TestPath = 'D:/Autism-Data/Kaggle/v' + str(Version) + '/test' TrainGen = keras.preprocessing.image.ImageDataGenerator( preprocessing_function=preprocess_input_new, horizontal_flip=True, rotation_range=45, width_shift_range=.01, height_shift_range=.01).flow_from_directory( TrainPath, target_size=(Height, Width), batch_size=BatchSize) ValidGen = keras.preprocessing.image.ImageDataGenerator( preprocessing_function=preprocess_input_new).flow_from_directory( ValidPath, target_size=(Height, Width), batch_size=BatchSize, shuffle=False) TestGen = keras.preprocessing.image.ImageDataGenerator( preprocessing_function=preprocess_input_new).flow_from_directory( TestPath, target_size=(Height, Width), batch_size=BatchSize, shuffle=False) os.makedirs("models/h5/" + str(timestr), exist_ok=True) filepath = "models/h5/" + str(timestr) + "/" + "weights-improvement-{epoch:02d}-{val_accuracy:.4f}.hdf5" SaveModelImage(model, "models/h5/" + str(timestr) + "/" + "Graph.png") copyfile('face.py', "models/h5/" + str(timestr) + "/face.py") checkpoint = keras.callbacks.callbacks.ModelCheckpoint(filepath, monitor='val_accuracy', verbose=1, save_best_only=True, mode='max') reduce_lr = keras.callbacks.callbacks.ReduceLROnPlateau(monitor='val_accuracy', factor=0.9, patience=5, min_lr=0.00001) ModelCallbacks = keras.callbacks.callbacks.LambdaCallback( on_epoch_begin=None, on_epoch_end=None, on_batch_begin=None, on_batch_end=None, on_train_begin=None, on_train_end=None) first = 5 if not load_model: # data = model.fit_generator( # generator = TrainGen, # validation_data= ValidGen, # epochs=first, # callbacks=[ModelCallbacks, reduce_lr, checkpoint], # verbose=1) models = [model] args = [{'generator':TrainGen, 'validation_data':TestGen, 'epochs':first, 'callbacks':[ModelCallbacks, reduce_lr, EarlyStoppingAtMinLoss()], 'verbose':1}] ml = MLEXPS() ml.setTopic('Autism') ml.setCopyFileList(['face.py']) ml.setModels(models) ml.setArgList(args) ml.generator = True ml.saveBestOnly = False ml.startExprQ() else: model = load_model(model_path) Y_pred = model.predict_generator(TestGen) y_pred = np.argmax(Y_pred, axis=1) print('Confusion Matrix') print(confusion_matrix(TestGen.classes, y_pred)) print('Classification Report') target_names = ['Autistic', 'Non_Autistic'] print(classification_report(TestGen.classes, y_pred, target_names=target_names))
[ "Bobar312@gmail.com" ]
Bobar312@gmail.com
b7b8ce02d0aba506b2683b3c8862f61ba4fd4293
9095c1a0da8c6ffe914ee6dd9c4708062fd95c9a
/vtpl_api/models/source_type.py
99b3143d277011d407f04a5955fab602b32550ca
[ "MIT" ]
permissive
vtpl1/vtpl_api_py
2e5338bd08677f12fc7304fb6ac7a32f32af1c93
d289c92254deb040de925205c583de69802a1c6b
refs/heads/master
2020-09-10T23:34:21.828350
2019-11-15T07:26:53
2019-11-15T07:26:53
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,504
py
# coding: utf-8 """ Engine api Engine APIs # noqa: E501 The version of the OpenAPI document: 1.0.4 Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six class SourceType(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ allowed enum values """ NONE = "none" RTSP = "rtsp" HTTP = "http" FILE = "file" FTP = "ftp" VMS = "vms" MQTT = "mqtt" AMQP = "amqp" S3 = "S3" VS3 = "VS3" BASEURL = "BaseUrl" RELATIVEURL = "RelativeUrl" """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { } attribute_map = { } def __init__(self): # noqa: E501 """SourceType - a model defined in OpenAPI""" # noqa: E501 self.discriminator = None def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, SourceType): return False return self.__dict__ == other.__dict__ def __ne__(self, other): """Returns true if both objects are not equal""" return not self == other
[ "monotosh.das@videonetics.com" ]
monotosh.das@videonetics.com
624821fbcf83feda5993b98a3444652ea2c64d7a
30d5f8094a42696be671917e8d1918363ccba963
/vd_design.py
7b255d16da74bad6bebd9e30f392359978bef17c
[]
no_license
ortariot/VMC-desktop
e8ffa6ba75c19bc5a785ed77da2bab6372a0878c
4df6b2a69f74cebf1c5b2cd722584fb1483a2b48
refs/heads/master
2023-05-31T15:56:56.304464
2021-06-11T12:04:20
2021-06-11T12:04:20
375,114,176
0
0
null
null
null
null
UTF-8
Python
false
false
12,886
py
# -*- coding: utf-8 -*- # Form implementation generated from reading ui file 'valve_design_desktop.ui' # # Created by: PyQt5 UI code generator 5.15.4 # # WARNING: Any manual changes made to this file will be lost when pyuic5 is # run again. Do not edit this file unless you know what you are doing. from PyQt5 import QtCore, QtGui, QtWidgets class Ui_MainWindow(object): def setupUi(self, MainWindow): MainWindow.setObjectName("MainWindow") MainWindow.setEnabled(True) MainWindow.resize(573, 475) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Fixed) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(MainWindow.sizePolicy( ).hasHeightForWidth()) MainWindow.setSizePolicy(sizePolicy) self.centralwidget = QtWidgets.QWidget(MainWindow) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Fixed) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.centralwidget.sizePolicy( ).hasHeightForWidth()) self.centralwidget.setSizePolicy(sizePolicy) self.centralwidget.setObjectName("centralwidget") self.horizontalLayout_5 = QtWidgets.QHBoxLayout(self.centralwidget) self.horizontalLayout_5.setObjectName("horizontalLayout_5") self.vlv1_frame = QtWidgets.QFrame(self.centralwidget) self.vlv1_frame.setEnabled(True) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.vlv1_frame.sizePolicy( ).hasHeightForWidth()) self.vlv1_frame.setSizePolicy(sizePolicy) self.vlv1_frame.setFrameShape(QtWidgets.QFrame.Panel) self.vlv1_frame.setFrameShadow(QtWidgets.QFrame.Raised) self.vlv1_frame.setObjectName("vlv1_frame") self.verticalLayout = QtWidgets.QVBoxLayout(self.vlv1_frame) self.verticalLayout.setObjectName("verticalLayout") self.vlv1_button = QtWidgets.QPushButton(self.vlv1_frame) self.vlv1_button.setObjectName("vlv1_button") self.verticalLayout.addWidget(self.vlv1_button) self.vlv1_lcdNumber = QtWidgets.QLCDNumber(self.vlv1_frame) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Minimum, QtWidgets.QSizePolicy.Fixed) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.vlv1_lcdNumber.sizePolicy( ).hasHeightForWidth()) self.vlv1_lcdNumber.setSizePolicy(sizePolicy) self.vlv1_lcdNumber.setMinimumSize(QtCore.QSize(0, 30)) font = QtGui.QFont() font.setPointSize(25) font.setBold(False) self.vlv1_lcdNumber.setFont(font) self.vlv1_lcdNumber.setObjectName("vlv1_lcdNumber") self.verticalLayout.addWidget(self.vlv1_lcdNumber) spacerItem = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout.addItem(spacerItem) self.widget = QtWidgets.QWidget(self.vlv1_frame) self.widget.setObjectName("widget") self.horizontalLayout = QtWidgets.QHBoxLayout(self.widget) self.horizontalLayout.setObjectName("horizontalLayout") self.vlv1_verticalSlider = QtWidgets.QSlider(self.widget) self.vlv1_verticalSlider.setEnabled(True) self.vlv1_verticalSlider.setOrientation(QtCore.Qt.Vertical) self.vlv1_verticalSlider.setObjectName("vlv1_verticalSlider") self.vlv1_verticalSlider.setEnabled(False) self.horizontalLayout.addWidget(self.vlv1_verticalSlider) self.verticalLayout.addWidget(self.widget) spacerItem1 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout.addItem(spacerItem1) self.horizontalLayout_5.addWidget(self.vlv1_frame) self.vlv2_frame = QtWidgets.QFrame(self.centralwidget) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.vlv2_frame.sizePolicy( ).hasHeightForWidth()) self.vlv2_frame.setSizePolicy(sizePolicy) self.vlv2_frame.setFrameShape(QtWidgets.QFrame.Panel) self.vlv2_frame.setFrameShadow(QtWidgets.QFrame.Raised) self.vlv2_frame.setObjectName("vlv2_frame") self.verticalLayout_2 = QtWidgets.QVBoxLayout(self.vlv2_frame) self.verticalLayout_2.setObjectName("verticalLayout_2") self.vlv2_button = QtWidgets.QPushButton(self.vlv2_frame) self.vlv2_button.setObjectName("vlv2_button") self.verticalLayout_2.addWidget(self.vlv2_button) self.vlv2_lcdNumber = QtWidgets.QLCDNumber(self.vlv2_frame) self.vlv2_lcdNumber.setMinimumSize(QtCore.QSize(0, 30)) self.vlv2_lcdNumber.setObjectName("vlv2_lcdNumber") self.verticalLayout_2.addWidget(self.vlv2_lcdNumber) spacerItem2 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_2.addItem(spacerItem2) self.widget1 = QtWidgets.QWidget(self.vlv2_frame) self.widget1.setObjectName("widget1") self.horizontalLayout_2 = QtWidgets.QHBoxLayout(self.widget1) self.horizontalLayout_2.setObjectName("horizontalLayout_2") self.vlv2_verticalSlider = QtWidgets.QSlider(self.widget1) self.vlv2_verticalSlider.setOrientation(QtCore.Qt.Vertical) self.vlv2_verticalSlider.setObjectName("vlv2_verticalSlider") self.vlv2_verticalSlider.setEnabled(False) self.horizontalLayout_2.addWidget(self.vlv2_verticalSlider) self.verticalLayout_2.addWidget(self.widget1) spacerItem3 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_2.addItem(spacerItem3) self.horizontalLayout_5.addWidget(self.vlv2_frame) self.vlv3_frame = QtWidgets.QFrame(self.centralwidget) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.vlv3_frame.sizePolicy( ).hasHeightForWidth()) self.vlv3_frame.setSizePolicy(sizePolicy) self.vlv3_frame.setFrameShape(QtWidgets.QFrame.Panel) self.vlv3_frame.setFrameShadow(QtWidgets.QFrame.Raised) self.vlv3_frame.setObjectName("vlv3_frame") self.verticalLayout_3 = QtWidgets.QVBoxLayout(self.vlv3_frame) self.verticalLayout_3.setObjectName("verticalLayout_3") self.vlv3_button = QtWidgets.QPushButton(self.vlv3_frame) self.vlv3_button.setObjectName("vlv3_button") self.verticalLayout_3.addWidget(self.vlv3_button) self.vlv3_lcdNumber = QtWidgets.QLCDNumber(self.vlv3_frame) self.vlv3_lcdNumber.setMinimumSize(QtCore.QSize(0, 30)) self.vlv3_lcdNumber.setObjectName("vlv3_lcdNumber") self.verticalLayout_3.addWidget(self.vlv3_lcdNumber) spacerItem4 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_3.addItem(spacerItem4) self.widget2 = QtWidgets.QWidget(self.vlv3_frame) self.widget2.setObjectName("widget2") self.horizontalLayout_3 = QtWidgets.QHBoxLayout(self.widget2) self.horizontalLayout_3.setObjectName("horizontalLayout_3") self.vlv3_verticalSlider = QtWidgets.QSlider(self.widget2) self.vlv3_verticalSlider.setOrientation(QtCore.Qt.Vertical) self.vlv3_verticalSlider.setObjectName("vlv3_verticalSlider") self.vlv3_verticalSlider.setEnabled(False) self.horizontalLayout_3.addWidget(self.vlv3_verticalSlider) self.verticalLayout_3.addWidget(self.widget2) spacerItem5 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_3.addItem(spacerItem5) self.horizontalLayout_5.addWidget(self.vlv3_frame) self.vlv4_frame = QtWidgets.QFrame(self.centralwidget) sizePolicy = QtWidgets.QSizePolicy(QtWidgets.QSizePolicy.Fixed, QtWidgets.QSizePolicy.Preferred) sizePolicy.setHorizontalStretch(0) sizePolicy.setVerticalStretch(0) sizePolicy.setHeightForWidth(self.vlv4_frame.sizePolicy( ).hasHeightForWidth()) self.vlv4_frame.setSizePolicy(sizePolicy) self.vlv4_frame.setFrameShape(QtWidgets.QFrame.Panel) self.vlv4_frame.setFrameShadow(QtWidgets.QFrame.Raised) self.vlv4_frame.setObjectName("vlv4_frame") self.verticalLayout_4 = QtWidgets.QVBoxLayout(self.vlv4_frame) self.verticalLayout_4.setObjectName("verticalLayout_4") self.vlv4_button = QtWidgets.QPushButton(self.vlv4_frame) self.vlv4_button.setObjectName("vlv4_button") self.verticalLayout_4.addWidget(self.vlv4_button) self.vlv4_lcdNumber = QtWidgets.QLCDNumber(self.vlv4_frame) self.vlv4_lcdNumber.setMinimumSize(QtCore.QSize(0, 30)) self.vlv4_lcdNumber.setObjectName("vlv4_lcdNumber") self.verticalLayout_4.addWidget(self.vlv4_lcdNumber) spacerItem6 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_4.addItem(spacerItem6) self.widget3 = QtWidgets.QWidget(self.vlv4_frame) self.widget3.setObjectName("widget3") self.horizontalLayout_4 = QtWidgets.QHBoxLayout(self.widget3) self.horizontalLayout_4.setObjectName("horizontalLayout_4") self.vlv4_verticalSlider = QtWidgets.QSlider(self.widget3) self.vlv4_verticalSlider.setOrientation(QtCore.Qt.Vertical) self.vlv4_verticalSlider.setObjectName("vlv4_verticalSlider") self.vlv4_verticalSlider.setEnabled(False) self.horizontalLayout_4.addWidget(self.vlv4_verticalSlider) self.verticalLayout_4.addWidget(self.widget3) spacerItem7 = QtWidgets.QSpacerItem(72, 20, QtWidgets.QSizePolicy.Expanding, QtWidgets.QSizePolicy.Minimum) self.verticalLayout_4.addItem(spacerItem7) self.horizontalLayout_5.addWidget(self.vlv4_frame) MainWindow.setCentralWidget(self.centralwidget) self.statusbar = QtWidgets.QStatusBar(MainWindow) self.statusbar.setObjectName("statusbar") MainWindow.setStatusBar(self.statusbar) self.retranslateUi(MainWindow) QtCore.QMetaObject.connectSlotsByName(MainWindow) def retranslateUi(self, MainWindow): _translate = QtCore.QCoreApplication.translate MainWindow.setWindowTitle(_translate("MainWindow", "Valve monster comander desktop")) self.vlv1_button.setText(_translate("MainWindow", "VALVE1 ON")) self.vlv2_button.setText(_translate("MainWindow", "VALVE2 ON")) self.vlv3_button.setText(_translate("MainWindow", "VALVE3 ON")) self.vlv4_button.setText(_translate("MainWindow", "VALVE4 ON")) if __name__ == "__main__": import sys app = QtWidgets.QApplication(sys.argv) MainWindow = QtWidgets.QMainWindow() ui = Ui_MainWindow() ui.setupUi(MainWindow) MainWindow.show() sys.exit(app.exec_())
[ "v.n.homutov@gmail.com" ]
v.n.homutov@gmail.com
50988401345a82502f2e380001be76f2af3666d2
25642cf53a941d9abac3f9b492ab76165e7ad4e9
/event-notification-server/sunBotApi/migrations/0006_event_closed.py
2e6fe9d95735a50ff509645afc96c2ab79d1117d
[]
no_license
Amwap/event-notification-bot
112f3be678d0e373a11b46b3b2342d92fa01bc75
8ea3f64cde606a161739f9bc0d1ece477f9a1cf5
refs/heads/master
2023-06-10T02:59:03.744669
2021-06-25T20:56:36
2021-06-25T20:56:36
380,349,377
0
0
null
null
null
null
UTF-8
Python
false
false
443
py
# Generated by Django 3.2.2 on 2021-05-14 23:02 from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('sunBotApi', '0005_auto_20210515_0101'), ] operations = [ migrations.AddField( model_name='event', name='closed', field=models.BooleanField(default=False, verbose_name='Закрыть приём заявок'), ), ]
[ "Amwap" ]
Amwap
27cd1801d257361237f2eacb2dbcb8e287f6685b
3f7d5999bb7e5a75454c8df2c5a8adcd1a8341ff
/tests/unit/modules/network/fortios/test_fortios_log_eventfilter.py
32a1e9c532163bad832b3009b0d154dc776ce8a7
[]
no_license
ansible-collection-migration/ansible.fortios
f7b1a7a0d4b69c832403bee9eb00d99f3be65e74
edad6448f7ff4da05a6c856b0e7e3becd0460f31
refs/heads/master
2020-12-18T13:08:46.739473
2020-02-03T22:10:49
2020-02-03T22:10:49
235,393,556
0
0
null
null
null
null
UTF-8
Python
false
false
8,342
py
# Copyright 2019 Fortinet, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <https://www.gnu.org/licenses/>. # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type import os import json import pytest from mock import ANY from ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios import FortiOSHandler try: from ansible_collections.ansible.fortios.plugins.modules import fortios_log_eventfilter except ImportError: pytest.skip("Could not load required modules for testing", allow_module_level=True) @pytest.fixture(autouse=True) def connection_mock(mocker): connection_class_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.modules.fortios_log_eventfilter.Connection') return connection_class_mock fos_instance = FortiOSHandler(connection_mock) def test_log_eventfilter_creation(mocker): schema_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.schema') set_method_result = {'status': 'success', 'http_method': 'POST', 'http_status': 200} set_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.set', return_value=set_method_result) input_data = { 'username': 'admin', 'state': 'present', 'log_eventfilter': { 'compliance_check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security_rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan_opt': 'enable', 'wireless_activity': 'enable' }, 'vdom': 'root'} is_error, changed, response = fortios_log_eventfilter.fortios_log(input_data, fos_instance) expected_data = { 'compliance-check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security-rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan-opt': 'enable', 'wireless-activity': 'enable' } set_method_mock.assert_called_with('log', 'eventfilter', data=expected_data, vdom='root') schema_method_mock.assert_not_called() assert not is_error assert changed assert response['status'] == 'success' assert response['http_status'] == 200 def test_log_eventfilter_creation_fails(mocker): schema_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.schema') set_method_result = {'status': 'error', 'http_method': 'POST', 'http_status': 500} set_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.set', return_value=set_method_result) input_data = { 'username': 'admin', 'state': 'present', 'log_eventfilter': { 'compliance_check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security_rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan_opt': 'enable', 'wireless_activity': 'enable' }, 'vdom': 'root'} is_error, changed, response = fortios_log_eventfilter.fortios_log(input_data, fos_instance) expected_data = { 'compliance-check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security-rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan-opt': 'enable', 'wireless-activity': 'enable' } set_method_mock.assert_called_with('log', 'eventfilter', data=expected_data, vdom='root') schema_method_mock.assert_not_called() assert is_error assert not changed assert response['status'] == 'error' assert response['http_status'] == 500 def test_log_eventfilter_idempotent(mocker): schema_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.schema') set_method_result = {'status': 'error', 'http_method': 'DELETE', 'http_status': 404} set_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.set', return_value=set_method_result) input_data = { 'username': 'admin', 'state': 'present', 'log_eventfilter': { 'compliance_check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security_rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan_opt': 'enable', 'wireless_activity': 'enable' }, 'vdom': 'root'} is_error, changed, response = fortios_log_eventfilter.fortios_log(input_data, fos_instance) expected_data = { 'compliance-check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security-rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan-opt': 'enable', 'wireless-activity': 'enable' } set_method_mock.assert_called_with('log', 'eventfilter', data=expected_data, vdom='root') schema_method_mock.assert_not_called() assert not is_error assert not changed assert response['status'] == 'error' assert response['http_status'] == 404 def test_log_eventfilter_filter_foreign_attributes(mocker): schema_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.schema') set_method_result = {'status': 'success', 'http_method': 'POST', 'http_status': 200} set_method_mock = mocker.patch('ansible_collections.ansible.fortios.plugins.module_utils.network.fortios.fortios.FortiOSHandler.set', return_value=set_method_result) input_data = { 'username': 'admin', 'state': 'present', 'log_eventfilter': { 'random_attribute_not_valid': 'tag', 'compliance_check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security_rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan_opt': 'enable', 'wireless_activity': 'enable' }, 'vdom': 'root'} is_error, changed, response = fortios_log_eventfilter.fortios_log(input_data, fos_instance) expected_data = { 'compliance-check': 'enable', 'endpoint': 'enable', 'event': 'enable', 'ha': 'enable', 'router': 'enable', 'security-rating': 'enable', 'system': 'enable', 'user': 'enable', 'vpn': 'enable', 'wan-opt': 'enable', 'wireless-activity': 'enable' } set_method_mock.assert_called_with('log', 'eventfilter', data=expected_data, vdom='root') schema_method_mock.assert_not_called() assert not is_error assert changed assert response['status'] == 'success' assert response['http_status'] == 200
[ "ansible_migration@example.com" ]
ansible_migration@example.com
4c4b3601e4073098253b9a1d67c3a5e373a569bd
546187161f8f0e234b4bf4e749bab477c32a352a
/arranging coins.py
fb508cafd1082fe4ee8b9ecf48efdc16b0ebb64d
[]
no_license
lzj322/leetcode
0fd71389c784715ce8adf975873a7e40d348b837
30ccbf3ae12d895f34c78e63c149c309468ab61b
refs/heads/master
2021-08-12T06:33:58.422303
2017-11-14T14:13:05
2017-11-14T14:13:05
110,697,832
0
0
null
null
null
null
UTF-8
Python
false
false
350
py
'LeetCode 441 arranging coins' import math class Solution(object): def arrangeCoins(self, n): """ :type n: int :rtype: int """ return int(((8*n+1)**0.5-1)/2) if __name__ == '__main__': s=Solution() l=[0,1,2,3,4,5,6,7,8,9] ans=list(map(s.arrangeCoins,l)) print (ans)
[ "lzj910322@gmail.com" ]
lzj910322@gmail.com
4312c5132af6818ca35ed0f704d81bfac2ddb825
5963c12367490ffc01c9905c028d1d5480078dec
/tests/components/wallbox/test_config_flow.py
6b5a05a3486830b64b8d0d53f7b409dfb288bb79
[ "Apache-2.0" ]
permissive
BenWoodford/home-assistant
eb03f73165d11935e8d6a9756272014267d7d66a
2fee32fce03bc49e86cf2e7b741a15621a97cce5
refs/heads/dev
2023-03-05T06:13:30.354545
2021-07-18T09:51:53
2021-07-18T09:51:53
117,122,037
11
6
Apache-2.0
2023-02-22T06:16:51
2018-01-11T16:10:19
Python
UTF-8
Python
false
false
5,028
py
"""Test the Wallbox config flow.""" import json from unittest.mock import patch import requests_mock from homeassistant import config_entries, data_entry_flow from homeassistant.components.wallbox import InvalidAuth, config_flow from homeassistant.components.wallbox.const import DOMAIN from homeassistant.core import HomeAssistant test_response = json.loads( '{"charging_power": 0,"max_available_power": 25,"charging_speed": 0,"added_range": 372,"added_energy": 44.697}' ) async def test_show_set_form(hass: HomeAssistant) -> None: """Test that the setup form is served.""" flow = config_flow.ConfigFlow() flow.hass = hass result = await flow.async_step_user(user_input=None) assert result["type"] == data_entry_flow.RESULT_TYPE_FORM assert result["step_id"] == "user" async def test_form_invalid_auth(hass): """Test we handle invalid auth.""" result = await hass.config_entries.flow.async_init( DOMAIN, context={"source": config_entries.SOURCE_USER} ) with patch( "homeassistant.components.wallbox.config_flow.WallboxHub.async_authenticate", side_effect=InvalidAuth, ): result2 = await hass.config_entries.flow.async_configure( result["flow_id"], { "station": "12345", "username": "test-username", "password": "test-password", }, ) assert result2["type"] == "form" assert result2["errors"] == {"base": "invalid_auth"} async def test_form_cannot_authenticate(hass): """Test we handle cannot connect error.""" result = await hass.config_entries.flow.async_init( DOMAIN, context={"source": config_entries.SOURCE_USER} ) with requests_mock.Mocker() as mock_request: mock_request.get( "https://api.wall-box.com/auth/token/user", text='{"jwt":"fakekeyhere","user_id":12345,"ttl":145656758,"error":false,"status":200}', status_code=403, ) mock_request.get( "https://api.wall-box.com/chargers/status/12345", text='{"Temperature": 100, "Location": "Toronto", "Datetime": "2020-07-23", "Units": "Celsius"}', status_code=403, ) result2 = await hass.config_entries.flow.async_configure( result["flow_id"], { "station": "12345", "username": "test-username", "password": "test-password", }, ) result2 = await hass.config_entries.flow.async_configure( result["flow_id"], { "station": "12345", "username": "test-username", "password": "test-password", }, ) assert result2["type"] == "form" assert result2["errors"] == {"base": "invalid_auth"} async def test_form_cannot_connect(hass): """Test we handle cannot connect error.""" result = await hass.config_entries.flow.async_init( DOMAIN, context={"source": config_entries.SOURCE_USER} ) with requests_mock.Mocker() as mock_request: mock_request.get( "https://api.wall-box.com/auth/token/user", text='{"jwt":"fakekeyhere","user_id":12345,"ttl":145656758,"error":false,"status":200}', status_code=200, ) mock_request.get( "https://api.wall-box.com/chargers/status/12345", text='{"Temperature": 100, "Location": "Toronto", "Datetime": "2020-07-23", "Units": "Celsius"}', status_code=404, ) result2 = await hass.config_entries.flow.async_configure( result["flow_id"], { "station": "12345", "username": "test-username", "password": "test-password", }, ) assert result2["type"] == "form" assert result2["errors"] == {"base": "cannot_connect"} async def test_form_validate_input(hass): """Test we handle cannot connect error.""" result = await hass.config_entries.flow.async_init( DOMAIN, context={"source": config_entries.SOURCE_USER} ) with requests_mock.Mocker() as mock_request: mock_request.get( "https://api.wall-box.com/auth/token/user", text='{"jwt":"fakekeyhere","user_id":12345,"ttl":145656758,"error":false,"status":200}', status_code=200, ) mock_request.get( "https://api.wall-box.com/chargers/status/12345", text='{"Temperature": 100, "Location": "Toronto", "Datetime": "2020-07-23", "Units": "Celsius"}', status_code=200, ) result2 = await hass.config_entries.flow.async_configure( result["flow_id"], { "station": "12345", "username": "test-username", "password": "test-password", }, ) assert result2["title"] == "Wallbox Portal" assert result2["data"]["station"] == "12345"
[ "noreply@github.com" ]
noreply@github.com
8946fdaa77811cae476e4d34aeade4fcaa751ef1
e200ad8a6e1aec6a3d3a55ed6223587a38460e27
/app.py
01e89251b10057eab1aca954f56b213b0fac7163
[]
no_license
user-cube/QRCode_Flask
3d1c09a3ba999c95430681d811f8acc6d2cc9eb6
6d8761005b3f0cf478f2ddbb383f2dafea0a92e8
refs/heads/master
2020-08-28T07:10:26.408452
2019-11-18T11:01:23
2019-11-18T11:01:23
217,631,168
0
0
null
null
null
null
UTF-8
Python
false
false
979
py
from flask import Flask, render_template, request, send_file from flask_qrcode import QRcode import home app = Flask(__name__) qrcode = QRcode(app) paths = home.Home.home(app) @app.route("/") @app.route('/home') def index(): return render_template("home.html", paths=paths) @app.route('/qrcode') def qrcodeGenerator(): return render_template("qrcode.html") @app.route('/qrcodeLogo') def qrcodeLogo(): return render_template("qrcodeLogo.html") @app.route("/genqrcode", methods=["GET"]) def get_qrcode(): data = request.args.get("data", "") return send_file(qrcode(data, mode="raw", box_size=100, error_correction='H'), mimetype="image/png") @app.route("/genqrcodeLogo", methods=["GET"]) def get_qrcodeImage(): data = request.args.get("data", "") image = request.args.get("image", "") return send_file(qrcode(data, mode="raw", box_size=100, error_correction='H', icon_img=image), mimetype="image/png") if __name__ == "__main__": app.run()
[ "ruicoelho@ua.pt" ]
ruicoelho@ua.pt
67dccdaf388e326388afec57b7acdf38c78908a9
eba0e40667d6082b5eeefdbaf2862e3f02fd774c
/mr_utils/sim/ssfp/quantitative_field_mapping.py
44a85af73a56bb265904c32bd1da3b6aaf216bbc
[]
no_license
zongjg/mr_utils
a0ec98ed2d03a6d52d81be8ef108993f92baeee1
08cb43dcf53fd6fddd3304e3514a608842310a34
refs/heads/master
2022-01-04T16:25:41.065177
2019-05-11T20:20:22
2019-05-11T20:20:22
null
0
0
null
null
null
null
UTF-8
Python
false
false
4,355
py
'''Quantitative field mapping for bSSFP. Collect quantitative MR maps (T1, T2, flip angle), then, assuming that these won't change during the duration of the scan, we can use these to take a single bSSFP scan each time point and solve for the off-resonance. Thus we get a field map at time point. ''' import numpy as np from mr_utils.utils import find_nearest from mr_utils.sim.ssfp import ssfp # from mr_utils import view def get_df_responses(T1, T2, PD, TR, alpha, phase_cyc, dfs): '''Simulate bSSFP response across all possible off-resonances. Parameters ========== T1 : float scalar T1 longitudinal recovery value in seconds. T2 : float scalar T2 transverse decay value in seconds. PD : float scalar proton density value scaled the same as acquisiton. TR : float Repetition time in seconds. alpha : float Flip angle in radians. phase_cyc : float RF phase cycling in radians. dfs : float Off-resonance values to simulate over. Returns ======= resp : array_like Frequency response of SSFP signal across entire spectrum. ''' # Feed ssfp sim an array of parameters to be used with all the df values T1s = np.ones(dfs.shape)*T1 T2s = np.ones(dfs.shape)*T2 PDs = np.ones(dfs.shape)*PD resp = ssfp(T1s, T2s, TR, alpha, dfs, phase_cyc=phase_cyc, M0=PDs) # Returns a vector of simulated Mxy with index corresponding to dfs return resp def quantitative_fm_scalar(Mxy, dfs, T1, T2, PD, TR, alpha, phase_cyc): '''For scalar T1, T2, PD. Parameters ========== Mxy : float Complex transverse signal we measure. dfs : array_like Off-resonance values to simulate over. T1 : float scalar T1 longitudinal recovery value in seconds. T2 : float scalar T2 transverse decay value in seconds. PD : float scalar proton density value scaled the same as acquisiton. TR : float Repetition time in seconds. alpha : float Flip angle in radians. phase_cyc : float RF phase cycling in radians. Returns ======= float Off-resonace value that most closely matches Mxy prior. ''' # Simulate over the total range of off-resonance values resp = get_df_responses(T1, T2, PD, TR, alpha, phase_cyc, dfs) # Find the response that matches Mxy most closely idx, _val = find_nearest(resp, Mxy) # Return the df's value, because that's really what the caller wanted return dfs[idx] def quantitative_fm(Mxys, dfs, T1s, T2s, PDs, TR, alpha, phase_cyc, mask=None): '''Find field map given quantitative maps. Parameters ========== Mxys : array_like Complex transverse signal we measure. dfs : array_like Off-resonance values to simulate over. T1s : array_like scalar T1 longitudinal recovery value in seconds. T2s : array_like scalar T2 transverse decay value in seconds. PDs : array_like scalar proton density value scaled the same as acquisiton. TR : float Repetition time in seconds. alpha : float Flip angle in radians. phase_cyc : float RF phase cycling in radians. mask : array_like Boolean mask to tell which pixels we should compute df for. Returns ======= fm : array_like Field map. ''' resps = {} orig_size = np.asarray(T1s).shape if mask is None: mask = np.ones(Mxys.shape) Mxys = np.asarray(Mxys).flatten() T1s = np.asarray(T1s).flatten() T2s = np.asarray(T2s).flatten() PDs = np.asarray(PDs).flatten() mask = np.asarray(mask).flatten() fm = np.zeros(Mxys.size) for ii in range(Mxys.size): if mask[ii]: # Cache results for later in case we come across the same T1,T2,PD if (PDs[ii], T1s[ii], T2s[ii]) not in resps: resps[(PDs[ii], T1s[ii], T2s[ii])] = get_df_responses( T1s[ii], T2s[ii], PDs[ii], TR, alpha, phase_cyc, dfs) # Find the appropriate off-resonance value for this T1,T2,PD,Mxy idx, _val = find_nearest( resps[(PDs[ii], T1s[ii], T2s[ii])], Mxys[ii]) fm[ii] = dfs[idx] else: fm[ii] = 0 return fm.reshape(orig_size)
[ "nicholas.bgp@gmail.com" ]
nicholas.bgp@gmail.com
397f63bed9a3a36120ebb0a445e6a6d487b5a736
670df3389c2dcad92e10f350dd40490eb6656f89
/src/networks/__init__.py
b7310380cf242e537dc0b4340da8ee7096d8bef3
[]
no_license
MoeinSorkhei/Thorax-Disease-Classification
8ed5494ef5db1ee86bf4eabf4ffdcf24eb80cf27
ef1f5f0289d62e132372974352bf974bf8b70e4c
refs/heads/master
2022-07-05T07:48:06.877423
2020-05-16T12:48:57
2020-05-16T12:48:57
261,700,037
0
0
null
null
null
null
UTF-8
Python
false
false
85
py
from .pre_trained import * from .trans_pool_pred import * from .init_models import *
[ "m.moein.sorkhei@gmail.com" ]
m.moein.sorkhei@gmail.com
6a7bd840b05232033b4479a414b2dba8cac470bb
d2fae2d0ff36fde8d8402bdac1de5b6760f050b7
/app/tests/Test_passwordchecker.py
031f23e09f40532aa833df7d554126e8cd5b2beb
[]
no_license
DennisMufasa/mongodb-flask_app
8701d817d757a5144b9a98ba4293a948c537b6c5
53c3447850d16d630428a020fe28949ff84c4a03
refs/heads/master
2022-12-09T11:31:59.085865
2020-08-31T02:32:57
2020-08-31T02:32:57
260,714,213
0
0
null
null
null
null
UTF-8
Python
false
false
484
py
# third-party import import unittest # local import from ..api.v1.models.utils import password_checker class Test_Password_checker(unittest.TestCase): def test_password_len(self): password_check1 = password_checker('boo') password_check2 = password_checker('lysergicaciddyethylammide') self.assertEqual(password_check1, 'password too short') self.assertEqual(password_check2, 'password too long') if __name__ == "__main__": unittest.main()
[ "denny.muasa@gmail.com" ]
denny.muasa@gmail.com
374f86075d5c187fad6bfde503fbdb0362a57e76
4985143dce9379c939d562d277350f0d8224f06a
/venv/bin/django-admin.py
e788bb50f4c0599e5161f6209905f79392df6d1e
[]
no_license
jkinathan/Task_todo
a74ae010dc703ba0ed4654a569b57a5ce7634857
e19da9ab9dede272b6c148b686e6e77e3da1687a
refs/heads/master
2023-03-23T13:51:41.816050
2021-03-20T09:32:32
2021-03-20T09:32:32
274,080,927
0
0
null
null
null
null
UTF-8
Python
false
false
162
py
#!/home/jo-kinathany/Desktop/Task_todo/venv/bin/python3 from django.core import management if __name__ == "__main__": management.execute_from_command_line()
[ "jkinobe@gmail.com" ]
jkinobe@gmail.com
5408d8eb40da578adce33eff9310ab27978446dd
8b0fdeee0f998f9d347789741251b447d8640e7e
/tf_encrypted/protocol/aby3/fp.py
633a7ae92f9472c0379b97e252111907d757e358
[ "LicenseRef-scancode-unknown-license-reference", "Apache-2.0" ]
permissive
tf-encrypted/tf-encrypted
639bd31190b7da7378de2c738e5e120309a4f027
2e7b569617697e49177e8b73c32957e1b939bbaf
refs/heads/master
2023-08-17T14:33:23.975918
2023-02-08T02:25:50
2023-02-08T02:25:50
126,222,784
1,007
196
Apache-2.0
2023-05-10T12:26:14
2018-03-21T18:22:13
Python
UTF-8
Python
false
false
11,638
py
# This file include the floating-point operations from math import ceil from math import log2 import numpy as np import tensorflow as tf from ..protocol import TFEPrivateTensor from ..protocol import TFEPublicTensor def _fp_div( prot, a: "TFEPrivateTensor", b: "TFEPrivateTensor", nonsigned: bool, precision: int = 1, ): with tf.name_scope("fp_div"): return a * _fp_recip_private(prot, b, nonsigned, precision=precision) def _fp_div_private_private( prot, a: "TFEPrivateTensor", b: "TFEPrivateTensor", nonsigned: bool, precision: int = 1, ): return _fp_div(prot, a, b, nonsigned, precision=precision) def _fp_div_public_private( prot, a: "TFEPublicTensor", b: "TFEPrivateTensor", nonsigned: bool, precision: int = 1, ): return _fp_div(prot, a, b, nonsigned, precision=precision) def _fp_div_private_public( prot, a: "TFEPrivateTensor", b: "TFEPublicTensor", nonsigned: bool, precision: int = 1, ): return a / b def _fp_div_public_public( prot, a: "TFEPublicTensor", b: "TFEPublicTensor", nonsigned: bool, precision: int = 1, ): return a / b def _fp_sqrt2(prot, a: "TFEPrivateTensor", precision: int = 1): c15 = prot.define_constant(1.5) c05 = prot.define_constant(0.5) y = approx_sqrt_inv(prot, a, precision=precision) # y approixmates 1 / sqrt(a) g = a * y h = y * c05 for _ in range(precision): """ Over iterations, g -> sqrt(b) h -> sqrt(1 / b) * 0.5 1 iteration should give less than 10^{-4} relative error """ r = c15 - g * h # r = 1 + error g = g * r h = h * r return g, h def _fp_recip_private(prot, x: "TFEPrivateTensor", nonsigned, precision: int = 1): """ Approxiamtedly compute 1/x from x. Apply the quintic iteration from http://numbers.computation.free.fr/Constants/Algorithms/inverse.html """ with tf.name_scope("fp_reciprocal"): sgf, exp = __fp_normalize( prot, x, nonsigned ) # x = sgf / exp, then 1 / x = 1/sgf * exp one = prot.define_constant(1.0) two = prot.define_constant(2, apply_scaling=False) # 2.9281928 - 2 * s approxiates 1/s with small relative errors # in the interval s \in [0.5, 1.) # By using integer factor '2', we can save one truncation. inv_sgf = 2.9281928 - two * sgf appr_recip = inv_sgf * exp # ~ 1/x for _ in range(precision): """One iteration should give us very good approximation (10^{-5} relative error ratio). More iterations, more precision.""" res = one - x * appr_recip res2 = res * res appr_recip = appr_recip + appr_recip * (one + res2) * ( res + res2 ) # quintic iteration return appr_recip def _fp_inv_sqrt_private(prot, a: "TFEPrivateTensor", precision: int = 1): # low precision return approx_sqrt_inv(prot, a, precision=precision) # high precision with extra 3 rounds of communication # two = prot.define_constant(2, apply_scaling=False) # _, h = _fp_sqrt2(prot, a) # return h * two def _fp_sqrt_private(prot, a: "TFEPrivateTensor", precision: int = 1): g, _ = _fp_sqrt2(prot, a, precision=precision) return g def prefix_ORs(b: "TFEPrivateTensor", k: int): r""" b := (..., b3, b2, b1, b0) where b0 is the least significant bit compute y := (y_{k-1}, y_{k-1}, ..., y0) where y_i \in {0, 1} y_i = bit_or_{i <= j < k} bj The first yi = 1 is the first significant bit that bi = 1. """ # running ORs from MSB to LSB. n, e = int(ceil(log2(k))), 1 with tf.name_scope("prefix_ORs"): for i in range(0, n): b = b | (b >> e) e = e << 1 return b def _do_fp_log_private(prot, x: "TFEPrivateTensor", base: "float", precision: int = 1): k = prot.fixedpoint_config.precision_fractional m = k + prot.fixedpoint_config.precision_integral n = prot.int_factory.nbits logn = ( int(log2(n) + 1) * 2 ) # enough bit length to represent the exponent \in [0, 128) assert x.is_scaled, "Er.. tricky here." assert ( 2 * k > m ), "We assume 2^{-j} can be represent with 2k-bit precisions for all j in [0, m)" assert base >= 2.0, "log(x, base) shoule with base >= 2." adjust = 1.0 / np.log2(base) with tf.name_scope("fp_log"): # bit-decomposition. Make sure the higher bits are all 0 via shifting. x_bits = (prot.a2b(x, m) << (n - m)) >> (n - m) y_bits = prefix_ORs(x_bits, m) z_bits = y_bits ^ (y_bits >> 1) rev_z_bits = prot.bit_reverse(z_bits) exponent = prot.b2a(rev_z_bits >> (n - 2 * k - 1), 2 * k) # NOTE: shift 1-less. exponent.is_scaled = True log_exponent = ( prot.b2a(prot.xor_indices(z_bits), logn) * 2**k ) # j + k with k-bit precision log_exponent.is_scaled = True log_exponent = log_exponent - k if base != 2.0: log_exponent = log_exponent * adjust frac = x * exponent # frac is in the interval [1., 2.] # The approximation coefficients are for log2(x), we need to # adjust them via multiplying 1./log2(base) """ """ log_frac = ((-0.4326728 * adjust) * frac + (2.276597 * adjust)) * frac + ( -1.843924 * adjust ) return log_frac + log_exponent def _fp_log2_private(prot, x: "TFEPrivateTensor"): return _do_fp_log_private(prot, x, 2.0) def _fp_log10_private(prot, x: "TFEPrivateTensor"): return _do_fp_log_private(prot, x, 10.0) def _fp_ln_private(prot, x: "TFEPrivateTensor"): return _do_fp_log_private(prot, x, np.e) def __fp_normalize(prot, b: "TFEPrivateTensor", nonsigned=False): r""" Given [b], to compute [sgf], and [exp] such that b = sgf / exp where sgf \in [0.5, 1) """ k = prot.fixedpoint_config.precision_fractional m = k + prot.fixedpoint_config.precision_integral n = b.backing_dtype.nbits assert b.is_scaled, "Er.. tricky here." assert ( 2 * k > m ), "We assume 2^{-j} can be represent with 2k-bit precisions for all j in [0, m)" with tf.name_scope("fp_normalize"): # juhou: sign.is_scaled is False if not nonsigned: msb = prot.bit_extract(b, m) two = np.ones(shape=msb.shape, dtype=int) * 2 sign = 1 - prot.mul_ab(prot.define_constant(two), msb) x = sign * b # abs(b) else: sign, x = 1, b # bit-decomposition. Make sure the higher bits are all 0 via shifting. x_bits = (prot.a2b(x, m) << (n - m)) >> (n - m) """ y_j = 1 <-> for all i <= j, y_i = 1. z_j = 1 <-> 2^j < x = 2^j + x0 < 2^{j+1}. And there is only one z_j = 1, and others are all 0. """ y_bits = prefix_ORs(x_bits, m) z = y_bits ^ (y_bits >> 1) r""" There exists one z_j = 1 and other z_i = 0. As as result, \sum_i 2^i * z_i = 2^j (i.e., b2a(z)) If we reverse the bits {z_i} to obtain {z'_i} for z'_i = z_{128 - i}. Then \sum_i 2^i * z'_i = 2^{128 - j}. Also, if we right shift {z'_i} by (128 - m)-steps. Then we obtain 2^{m - j}, which equals to the fixed-point representation of 2^{-j} within m-bits precision. """ # NOTE: We couldn't obtain 2^{-j} with only k-bits precision # because j > k might be possible (i.e., b = 2^k * b' for # real value b' > 1). # Instead, we preserve 2k-bit precision, i.e., c = 2^{-j} with 2k bits precision _exp = prot.b2a(prot.bit_reverse(z) >> (n - 2 * k), 2 * k) _exp.is_scaled = True sgf = x * _exp # significant should in [0.5, 1). Thus, sgf should be positive. exp = _exp if nonsigned else sign * _exp return sgf, exp def approx_sqrt_inv(prot, x: "TFEPrivateTensor", precision: int = 1): """ From x, to compute an approximation of 1/sqrt(x). """ def select(x, y, bit): """ return y if bit = 0 else x. """ c = np.ones(shape=bit.shape) * (x - y) return prot.mul_ab(prot.define_constant(c), bit) + y k = prot.fixedpoint_config.precision_fractional n = x.backing_dtype.nbits # using top half bits as integer, bottom half bits as fraction. s = (n // 2) - k xs = x << s assert k >= s assert n % 2 == 0 assert x.is_scaled, "Er.. tricky here." with tf.name_scope("inv_sqrt"): # bit-decomposition. x_bits = prot.a2b(xs, n) y_bits = prefix_ORs(x_bits, n) z_bits = ( y_bits ^ (y_bits >> 1) ) << 1 # note: x = c * 2^m where c \in [0.25, 0.5) rev_z_bits = prot.bit_reverse(z_bits) frac = prot.b2a(rev_z_bits, n) # By default, bottom k bits used as fraction, # do truncate manually here. frac.is_scaled = False normalized = frac * x # normalized \in [0.25, 0.5) normalized = prot.truncate(normalized, amount=k + s) normalized.is_scaled = True """ f(b) = 4.7979 * b^2 - 5.9417 * b + 3.1855 approixmates 1/sqrt(b) in [0.25, 0.5) with less than 0.7% relative error """ sqrt_inv = ((4.7979 * normalized) - 5.9417) * normalized + 3.1855 # more iteration, more precision for i in range(precision): sqrt_inv = sqrt_inv * (3 - normalized * sqrt_inv * sqrt_inv) sqrt_inv = sqrt_inv * 0.5 """ Indeed, the exponetent part is 2^{j+k+s} where k is the scaling factor, s = (n // 2) - k We want to compute sqrt(2^{-j}) with k-bit precision, i.e., sqrt(2^{-j}) * 2^k. In other words, we compute sqrt(2^{-j}) * 2^k from 2^{j+k+s}. 1. We first obtain 2^{-(j+k+s)} from 2^{j+k+s}. 2. Then we compute 2^{floor(-(j+k+s)/2)}. Rewrite it as 2^{floor(-(j+k+s)/2)} = c * 2^{floor(-j/2)} * 2^{floor(-(k+s)/2)} where c depends on the parity of j, and k + s. 3. We compute the parity of j+k+s, i.e., check the LSB of j+k+s. 4. Suppose k+s is even, 2^{floor((s-k)/2)} = 2^{(s-k)/2}. Then we can cancel this term via 2^{s-k/2}. If lsb(j+k+s) = 0 <-> j is even. In this case, 2^{floor(-j/2)} = 2^{-j/2} = sqrt(2^{-j}). If lsb(j+k+s) = 1 <-> j is odd. Then 2^{floor(-j/2)} * 2^{-1/2} = sqrt(2^{-j}). Suppose k+s is odd, We need 2^{(s-k)//2} * 2 to cancel 2^{floor(-(k+s)/2)}. If lsb(j+k+s) = 0 <-> j is odd. In this case, 2^{floor(-j/2)} * 2^{-1/2} = sqrt(2^{-j}). If lsb(j+k+s) = 1 <-> j is even. Then 2^{floor(-j/2)} = 2^{-j/2} = sqrt(2^{-j}). """ sum_jks = prot.xor_indices(z_bits) lsb = prot.bit_extract(sum_jks, 0) # lsb = 0 <-> j+k+s is even exponet = prot.b2a( prot.bit_gather(rev_z_bits | rev_z_bits >> 1, 0, 2), k + s ) # 2^{floor(-(j+k+s)/2)} esk = 2 ** ((k - s) // 2) if (k + s) & 1 == 0: # k+s is even which means lsb = 1 <=> j is odd exponet = exponet * select(esk, esk * np.sqrt(2.0), lsb) else: # k+s is odd which means lsb = 1 <=> j is even exponet = exponet * select(esk * np.sqrt(2.0), esk * 2, lsb) return sqrt_inv * exponet
[ "noreply@github.com" ]
noreply@github.com
4eacd86775bb8afb4f12c25a8a0907c0cc30960e
0a674e831b4616d8013511b88c1097c29474042d
/tutorials/tf_modularity.py
294fc17e29699ea9f545f741077b3c652f180045
[]
no_license
ncullen93/tensorflow_portfolio
eac119cba6ffd69146326828dccfc635daff60c0
b14145dc7d8cd2fe86e036033d2890548ae09456
refs/heads/master
2021-03-19T12:22:17.617012
2017-10-14T16:42:14
2017-10-14T16:42:14
86,193,597
2
1
null
null
null
null
UTF-8
Python
false
false
524
py
""" Code for reusing common tensorflow functionality """ def relu(X, n_out=1): with tf.name_scope('relu'): w_shape = (int(X.get_shape()[1]),n_out) w = tf.Variable(tf.random_normal(w_shape), name='weights') b = tf.Variable(0.0, name='bias') z = tf.add(tf.matmul(X,w),b, name='z') return tf.nn.relu(z, name='relu') n_features = 3 X = tf.placeholder(tf.float32, shape=(None,n_features), name='X') relus = [relu(X) for i in range(5)] relu_addition = tf.add_n(relus, name='output')
[ "ncullen@Nicks-MacBook-Pro.local" ]
ncullen@Nicks-MacBook-Pro.local
0eb1cfa23de5563d2ffe22de91ace92cb6fa8e27
9e98b6ce7b6bc8a521a07196a43ccf4b7f697f78
/env/bin/easy_install
2bba1c50de5452763acaaf86a978462ffde83587
[]
no_license
gomeztagle-alan/lab2
53f25a20281b3a46cac164fbe5b16c217f5beb25
84e80e053160f6d9d818cc85f4274f848e62be6a
refs/heads/master
2023-02-10T10:54:12.073327
2019-07-16T22:54:46
2019-07-16T22:54:46
194,698,854
0
0
null
null
null
null
UTF-8
Python
false
false
269
#!/home/arcu5/Desktop/e14a/week1/lab2/env/bin/python2 # -*- coding: utf-8 -*- import re import sys from setuptools.command.easy_install import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "gosanocubed@gmail.com" ]
gosanocubed@gmail.com
faa2e47e01b26f98eb24501a23c59d2dd2f3081a
70bc77336e4544031ad7d7d29a2e964ef2626076
/base/models.py
bf4ba34fec4fc78262b81397124b4041d26e64fd
[]
no_license
DronMDF/vanadis
9af7a8c9281bf0eb17df593f5c9fc9345e474612
de692207bbd127c5a9952e3144653492a0ba969f
refs/heads/master
2020-04-17T08:11:18.411429
2016-12-21T20:50:05
2016-12-21T20:50:05
66,539,179
1
0
null
2016-12-21T20:50:06
2016-08-25T08:20:03
Python
UTF-8
Python
false
false
654
py
from django.db import models class Project(models.Model): name = models.CharField(max_length=100, db_index=True) repo_url = models.CharField(max_length=256, null=True) class Object(models.Model): project = models.ForeignKey(Project, on_delete=models.CASCADE, db_index=True) oid = models.BigIntegerField(db_index=True) issues_count = models.IntegerField() class Issue(models.Model): project = models.ForeignKey(Project, on_delete=models.CASCADE, db_index=True) object = models.ForeignKey(Object, on_delete=models.CASCADE, db_index=True) line = models.IntegerField() position = models.IntegerField() text = models.CharField(max_length=256)
[ "dron.valyaev@gmail.com" ]
dron.valyaev@gmail.com
a2d97895a54972443c74c32f10d1ee5f07775db6
2de1aaa4d9bde03e21601bfeddf7e85863b0d54a
/django/jason_test/booktest/views.py
1592057bc085ff5a7c9206432df946588d5f918e
[]
no_license
jasondzy/Python
d0a0ffad8aaac1f75d9e7c5ecd5cc9c33bd712ea
9fe9cf23f7defa3581e7bcfe4cf8ec6a830b6cd7
refs/heads/master
2021-01-01T18:46:00.739387
2018-02-11T14:59:26
2018-02-11T14:59:26
98,430,631
0
0
null
null
null
null
UTF-8
Python
false
false
2,706
py
from django.shortcuts import render from django.http import HttpResponse, JsonResponse from django.conf import settings from .models import * from django.core.paginator import Paginator import json # Create your views here. def index(request): return render(request,'booktest/index.html') # return HttpResponse('ok') def upload(request): if request.method == 'POST': picture = request.FILES['picture'] fname = '%s/cars/%s'%(settings.MEDIA_ROOT, picture.name) # return HttpResponse(settings.MEDIA_ROOT) with open(fname,'wb') as pic: for c in picture.chunks(): pic.write(c) return HttpResponse('ok') else: return HttpResponse('error') def my_custom_page_not_found_view(request): return HttpResponse('fail 404 fail') def pages(request,id): if id == '': id = '1' list = HeroInfo.objects.all() paginator = Paginator(list,5) page = paginator.page(int(id)) context = {'page':page} return render(request,'booktest/pages.html',context) def book_index(request): return render(request,'booktest/book_index.html') def ajax_get(request): # print('hello') book_list = BookInfo.objects.all() # print(book_list) # return JsonResponse({'data':book_list}) #这样返回数据会报错 l = [] for list in book_list: l.append((list.id,list.btitle)) #这里必须要将获得的book_list进行遍历,取出元素放在一个数组中才行,否则会报错 return JsonResponse({'data':l}) #具体为什么这样的原因还有待解析 def get_bookinfo(request): print('test') id = request.GET['id'] #这里使用的是ajax的方式进行数据的传递 print(id) list=HeroInfo.objects.filter(hBook_id=id)#这里的filter返回的是一个查询集,并不是一个对象,而是一个对象集合 print(list) hero_list = [] for i in list: hero_list.append((i.id,i.hname)) return JsonResponse({'data':hero_list}) #富文本编辑器 def editor(request): data = BookInfo.objects.all() list = [] for l in data: list.append([l.id,l.btitle]) print(list) context = {'data':list} return render(request,'booktest/editor.html',context) def editor_handle(request): html = request.POST['hcontent'] #此处获取的是大文本提交的内容 id = request.POST['select'] #此处获取的是要提交的id号 print('id:',id) print(html) book = BookInfo.objects.get(pk=id) #注意filter方法返回的是一个查询集合,是一个集合,get返回的是一个对象 print(book.btitle) book.bcontent = html book.save() return HttpResponse('ok')
[ "812724347@qq.com" ]
812724347@qq.com
a0f7ba2af8f27035c0405bb59bf919df9026b98a
ffa667819d306d634b15d089f5cef9dbaf652abb
/longan_sqlite/util.py
f5bf9740129cb819b172f44add348691f7be5d2b
[ "MIT" ]
permissive
xiaobaiso/longan-sqlite3
084a08e5e5e70b165fcc60da224cb2bbe6c761b8
6a260c913f96bdf90b7e0f6bb3f9d8f1a57e9f57
refs/heads/master
2020-03-18T08:54:50.466931
2018-05-21T04:44:42
2018-05-21T04:44:42
null
0
0
null
null
null
null
UTF-8
Python
false
false
468
py
from .flesh import Flesh def convert_dicts(fields, items): """ :param fields: :param items: :return: list(Flesh) :rtype: list """ ret_items = [] for i in items: item_dict = {} for k, v in enumerate(fields): item_dict[v] = i[k] ret_items.append(Flesh(item_dict)) return ret_items def add_quotes(value): if isinstance(value, str): value = '"{}"'.format(value) return str(value)
[ "yaoma@58ganji.com" ]
yaoma@58ganji.com
e51a73b1b803cf62864be83be78a2bba41b8c063
5101bfc97d22e8e4465ba9f86779a9b64a469d63
/U-Net_for_Retinal Vessel Segmentation/train.py
866900ce43cd92686fa0df6ec9c4d82354d15773
[]
no_license
WangMeow1998/Medical-Image-Segmentation
44214aedbefc42490a402e37491350531603a6a7
719220dc2471e1de42f8d1823295c3ef2cd35ce4
refs/heads/master
2023-02-02T10:29:21.520466
2020-12-24T03:51:06
2020-12-24T03:51:06
315,533,215
0
0
null
null
null
null
UTF-8
Python
false
false
3,539
py
from unet_model import UNet from utils.DALOADER import DealDataset from torch import optim import torch.nn as nn import torch from tqdm import tqdm import sys class Logger(): def __init__(self, filename="log.txt"): self.terminal = sys.stdout self.log = open(filename, "w") def write(self, message): self.terminal.write(message) self.log.write(message) def flush(self): pass class SoftDiceLoss(nn.Module): def __init__(self, weight=None, size_average=True): super(SoftDiceLoss, self).__init__() def forward(self, logits, targets): num = targets.size(0) smooth = 1 probs = torch.sigmoid(logits) m1 = probs.view(num, -1) m2 = targets.view(num, -1) intersection = (m1 * m2) score = 2. * (intersection.sum(1) + smooth) / (m1.sum(1) + m2.sum(1) + smooth) score = 1 - score.sum() / num return score def dice_coeff(pred, target): smooth = 1. num = pred.size(0) m1 = pred.view(num, -1) # Flatten m2 = target.view(num, -1) # Flatten intersection = (m1 * m2).sum() return (2. * intersection + smooth) / (m1.sum() + m2.sum() + smooth) def train_net(net, device, data_path, epochs=300, batch_size=2, lr=0.0001): # 加载训练集 name_dataset = DealDataset(data_path) train_loader = torch.utils.data.DataLoader(dataset=name_dataset, batch_size=batch_size, shuffle=True) #shuffle 填True 就会打乱 #定义算法 optimizer = optim.Adam(net.parameters(), lr=lr, weight_decay=1e-8) # optimizer = optim.SGD(net.parameters(), lr=lr, momentum=0.9) # scheduler = optim.lr_scheduler.StepLR(optimizer, step_size=100, gamma=0.3, last_epoch=-1) #定义Loss loss1 = nn.BCEWithLogitsLoss() loss2 = SoftDiceLoss() # best_loss统计,初始化为正无穷 # best_loss = float('inf') # best_dice = -best_loss sys.stdout = Logger() for epoch in tqdm(range(epochs)): # 训练模式 net.train() for batch in train_loader: optimizer.zero_grad() image = batch['image'] label = batch['label'] image = image.to(device=device, dtype=torch.float32) label = label.to(device=device, dtype=torch.float32) pred = net(image) # 计算loss c1 = loss1(pred, label) c2 = loss2(pred, label) loss = c1+c2 # loss = c2 pred = torch.sigmoid(pred) pred = (pred>0.5).float() print(pred.sum()) dice = dice_coeff(pred, label) print('Dice/train', dice.item(),'\t','Loss/train', loss.item()) # 保存loss值最小的网络参数 # if dice > best_dice: # best_dice = dice # torch.save(net.state_dict(), 'best_model2021.pth') # 更新参数 loss.backward() optimizer.step() torch.save(net.state_dict(), f'CP_epoch{epoch + 1}.pth') # scheduler.step() if __name__ == "__main__": # 选择设备,有cuda用cuda,没有就用cpu device = torch.device('cuda' if torch.cuda.is_available() else 'cpu') # 加载网络,图片单通道1,分类为1。 net = UNet(n_channels=3, n_classes=1) # 将网络拷贝到deivce中 net.to(device=device) # 指定训练集地址,开始训练 # data_path = "../input/my-eye/data/CHASE/train" data_path = "./data/CHASE/train" train_net(net, device, data_path)
[ "wangmeow1998@163.com" ]
wangmeow1998@163.com
e884b3995dc86624872f534df6a05bcb1cc306c6
e5e55e853dc834177f03cf8bb5161e847baa9830
/consts.py
807ce52fe670cf0645cb541a4bae87c9750e1bc0
[ "MIT" ]
permissive
mmmdamin/mastermind
2b62864443db1c9bc2dab47d5aed818f6cd0e991
d775e0492165acba8951fe308330f38bf88947d6
refs/heads/master
2021-09-02T07:34:47.015269
2017-12-31T14:29:10
2017-12-31T14:29:10
115,865,993
2
1
null
null
null
null
UTF-8
Python
false
false
256
py
GAME_LENGTH = 4 COLORS = { 1: "red", 2: "blue", 3: "green", 4: "yellow", 5: "white", } MIN_COLOR = min(COLORS.keys()) MAX_COLOR = max(COLORS.keys()) MAX_TURNS = 1 WIN_MESSAGE = "You won!" LOSE_MESSAGE = "You lose! initial state {}"
[ "sabbaghian@arsh.co" ]
sabbaghian@arsh.co
de2b6b74989a2467127597423d029e5b5810eb06
f576f0ea3725d54bd2551883901b25b863fe6688
/sdk/compute/azure-mgmt-vmwarecloudsimple/generated_samples/list_customization_policies.py
7deb911d492ae31425baf1a5011cba20636db4d2
[ "LicenseRef-scancode-generic-cla", "MIT", "LGPL-2.1-or-later" ]
permissive
Azure/azure-sdk-for-python
02e3838e53a33d8ba27e9bcc22bd84e790e4ca7c
c2ca191e736bb06bfbbbc9493e8325763ba990bb
refs/heads/main
2023-09-06T09:30:13.135012
2023-09-06T01:08:06
2023-09-06T01:08:06
4,127,088
4,046
2,755
MIT
2023-09-14T21:48:49
2012-04-24T16:46:12
Python
UTF-8
Python
false
false
1,604
py
# coding=utf-8 # -------------------------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # Licensed under the MIT License. See License.txt in the project root for license information. # Code generated by Microsoft (R) AutoRest Code Generator. # Changes may cause incorrect behavior and will be lost if the code is regenerated. # -------------------------------------------------------------------------- from azure.identity import DefaultAzureCredential from azure.mgmt.vmwarecloudsimple import VMwareCloudSimple """ # PREREQUISITES pip install azure-identity pip install azure-mgmt-vmwarecloudsimple # USAGE python list_customization_policies.py Before run the sample, please set the values of the client ID, tenant ID and client secret of the AAD application as environment variables: AZURE_CLIENT_ID, AZURE_TENANT_ID, AZURE_CLIENT_SECRET. For more info about how to get the value, please see: https://docs.microsoft.com/azure/active-directory/develop/howto-create-service-principal-portal """ def main(): client = VMwareCloudSimple( credential=DefaultAzureCredential(), subscription_id="{subscription-id}", ) response = client.customization_policies.list( region_id="myResourceGroup", pc_name="myPrivateCloud", ) for item in response: print(item) # x-ms-original-file: specification/vmwarecloudsimple/resource-manager/Microsoft.VMwareCloudSimple/stable/2019-04-01/examples/ListCustomizationPolicies.json if __name__ == "__main__": main()
[ "noreply@github.com" ]
noreply@github.com
7fc429a82b9d8ab58b2b4f3d058c32c644b665b4
c43ab8680e571136899564070928bfa9d54612d3
/addons/io_scene_gltf2/__init__.py
2ea8edd3846ad4a2159f02ddd41824637c83729a
[ "Apache-2.0" ]
permissive
cuulee/glTF-Blender-IO
447132a934a381b736dc4eaa1cc885d2d777c7b0
fb1bc8dbaffb416502a2ac713b1fe3ad1d4c481e
refs/heads/master
2020-03-26T01:53:11.525100
2018-08-11T07:33:26
2018-08-11T07:33:26
null
0
0
null
null
null
null
UTF-8
Python
false
false
16,859
py
# Copyright (c) 2018 The Khronos Group Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # # Imports # import bpy import os from bpy_extras.io_utils import ImportHelper from bpy.types import Operator from .blender.imp.io import * from .blender.imp.scene import * from .blender.imp.util import * from bpy.props import (CollectionProperty, StringProperty, BoolProperty, EnumProperty, FloatProperty, IntProperty) from bpy_extras.io_utils import (ExportHelper) # # Globals # bl_info = { 'name': 'glTF 2.0 format', 'author': 'Julien Duroure & Norbert Nopper', 'blender': (2, 79, 0), 'location': 'File > Import-Export', 'description': 'Import-Export as glTF 2.0', 'warning': '', 'wiki_url': '' '', 'support': 'COMMUNITY', 'category': 'Import-Export'} # # Functions / Classes. # class GLTF2ExportSettings(bpy.types.Operator): """Save the export settings on export (saved in .blend). Toggle off to clear settings""" bl_label = "Save Settings" bl_idname = "scene.gltf2_export_settings_set" def execute(self, context): operator = context.active_operator operator.will_save_settings = not operator.will_save_settings if not operator.will_save_settings and context.scene.get(operator.scene_key, False): # clear settings del context.scene[operator.scene_key] return {"FINISHED"} class ExportGLTF2_Base(): export_copyright = StringProperty( name='Copyright', description='', default='' ) export_embed_buffers = BoolProperty( name='Embed buffers', description='', default=False ) export_embed_images = BoolProperty( name='Embed images', description='', default=False ) export_strip = BoolProperty( name='Strip delimiters', description='', default=False ) export_indices = EnumProperty( name='Maximum indices', items=(('UNSIGNED_BYTE', 'Unsigned Byte', ''), ('UNSIGNED_SHORT', 'Unsigned Short', ''), ('UNSIGNED_INT', 'Unsigned Integer', '')), default='UNSIGNED_INT' ) export_force_indices = BoolProperty( name='Force maximum indices', description='', default=False ) export_texcoords = BoolProperty( name='Export texture coordinates', description='', default=True ) export_normals = BoolProperty( name='Export normals', description='', default=True ) export_tangents = BoolProperty( name='Export tangents', description='', default=True ) export_materials = BoolProperty( name='Export materials', description='', default=True ) export_colors = BoolProperty( name='Export colors', description='', default=True ) export_cameras = BoolProperty( name='Export cameras', description='', default=False ) export_camera_infinite = BoolProperty( name='Infinite perspective Camera', description='', default=False ) export_selected = BoolProperty( name='Export selected only', description='', default=False ) export_layers = BoolProperty( name='Export all layers', description='', default=True ) export_extras = BoolProperty( name='Export extras', description='', default=False ) export_yup = BoolProperty( name='Convert Z up to Y up', description='', default=True ) export_apply = BoolProperty( name='Apply modifiers', description='', default=False ) export_animations = BoolProperty( name='Export animations', description='', default=True ) export_frame_range = BoolProperty( name='Export within playback range', description='', default=True ) export_frame_step = IntProperty( name='Frame step size', description='Step size (in frames) for animation export.', default=1, min=1, max=120 ) export_move_keyframes = BoolProperty( name='Keyframes start with 0', description='', default=True ) export_force_sampling = BoolProperty( name='Force sample animations', description='', default=False ) export_current_frame = BoolProperty( name='Export current frame', description='', default=True ) export_skins = BoolProperty( name='Export skinning', description='', default=True ) export_bake_skins = BoolProperty( name='Bake skinning constraints', description='', default=False ) export_morph = BoolProperty( name='Export morphing', description='', default=True ) export_morph_normal = BoolProperty( name='Export morphing normals', description='', default=True ) export_morph_tangent = BoolProperty( name='Export morphing tangents', description='', default=True ) export_lights = BoolProperty( name='Export KHR_lights_punctual', description='', default=False ) export_displacement = BoolProperty( name='Export KHR_materials_displacement', description='', default=False ) will_save_settings = BoolProperty(default=False) # Custom scene property for saving settings scene_key = "glTF2ExportSettings" # def invoke(self, context, event): settings = context.scene.get(self.scene_key) self.will_save_settings = False if settings: try: for (k,v) in settings.items(): setattr(self, k, v) self.will_save_settings = True except AttributeError: self.report({"ERROR"}, "Loading export settings failed. Removed corrupted settings") del context.scene[self.scene_key] return ExportHelper.invoke(self, context, event) def save_settings(self, context): # find all export_ props all_props = self.properties export_props = {x:all_props.get(x) for x in dir(all_props) if x.startswith("export_") and all_props.get(x) is not None} context.scene[self.scene_key] = export_props def execute(self, context): from .blender.exp import gltf2_blender_export if self.will_save_settings: self.save_settings(context) # All custom export settings are stored in this container. export_settings = {} export_settings['gltf_filepath'] = bpy.path.ensure_ext(self.filepath, self.filename_ext) export_settings['gltf_filedirectory'] = os.path.dirname(export_settings['gltf_filepath']) + '/' export_settings['gltf_format'] = self.export_format export_settings['gltf_copyright'] = self.export_copyright export_settings['gltf_embed_buffers'] = self.export_embed_buffers export_settings['gltf_embed_images'] = self.export_embed_images export_settings['gltf_strip'] = self.export_strip export_settings['gltf_indices'] = self.export_indices export_settings['gltf_force_indices'] = self.export_force_indices export_settings['gltf_texcoords'] = self.export_texcoords export_settings['gltf_normals'] = self.export_normals export_settings['gltf_tangents'] = self.export_tangents and self.export_normals export_settings['gltf_materials'] = self.export_materials export_settings['gltf_colors'] = self.export_colors export_settings['gltf_cameras'] = self.export_cameras if self.export_cameras: export_settings['gltf_camera_infinite'] = self.export_camera_infinite else: export_settings['gltf_camera_infinite'] = False export_settings['gltf_selected'] = self.export_selected export_settings['gltf_layers'] = self.export_layers export_settings['gltf_extras'] = self.export_extras export_settings['gltf_yup'] = self.export_yup export_settings['gltf_apply'] = self.export_apply export_settings['gltf_animations'] = self.export_animations if self.export_animations: export_settings['gltf_current_frame'] = False export_settings['gltf_frame_range'] = self.export_frame_range export_settings['gltf_move_keyframes'] = self.export_move_keyframes export_settings['gltf_force_sampling'] = self.export_force_sampling else: export_settings['gltf_current_frame'] = self.export_current_frame export_settings['gltf_frame_range'] = False export_settings['gltf_move_keyframes'] = False export_settings['gltf_force_sampling'] = False export_settings['gltf_skins'] = self.export_skins if self.export_skins: export_settings['gltf_bake_skins'] = self.export_bake_skins else: export_settings['gltf_bake_skins'] = False export_settings['gltf_frame_step'] = self.export_frame_step export_settings['gltf_morph'] = self.export_morph if self.export_morph: export_settings['gltf_morph_normal'] = self.export_morph_normal else: export_settings['gltf_morph_normal'] = False if self.export_morph and self.export_morph_normal: export_settings['gltf_morph_tangent'] = self.export_morph_tangent else: export_settings['gltf_morph_tangent'] = False export_settings['gltf_lights'] = self.export_lights export_settings['gltf_displacement'] = self.export_displacement export_settings['gltf_binary'] = bytearray() export_settings['gltf_binaryfilename'] = os.path.splitext(os.path.basename(self.filepath))[0] + '.bin' return gltf2_blender_export.save(self, context, export_settings) def draw(self, context): layout = self.layout # col = layout.box().column() col.label('Embedding:', icon='PACKAGE') col.prop(self, 'export_copyright') if self.export_format == 'ASCII': col.prop(self, 'export_embed_buffers') col.prop(self, 'export_embed_images') col.prop(self, 'export_strip') col = layout.box().column() col.label('Nodes:', icon='OOPS') col.prop(self, 'export_selected') col.prop(self, 'export_layers') col.prop(self, 'export_extras') col.prop(self, 'export_yup') col = layout.box().column() col.label('Meshes:', icon='MESH_DATA') col.prop(self, 'export_apply') col.prop(self, 'export_indices') col.prop(self, 'export_force_indices') col = layout.box().column() col.label('Attributes:', icon='SURFACE_DATA') col.prop(self, 'export_texcoords') col.prop(self, 'export_normals') if self.export_normals: col.prop(self, 'export_tangents') col.prop(self, 'export_colors') col = layout.box().column() col.label('Objects:', icon='OBJECT_DATA') col.prop(self, 'export_cameras') if self.export_cameras: col.prop(self, 'export_camera_infinite') col = layout.box().column() col.label('Materials:', icon='MATERIAL_DATA') col.prop(self, 'export_materials') col = layout.box().column() col.label('Animation:', icon='OUTLINER_DATA_POSE') col.prop(self, 'export_animations') if self.export_animations: col.prop(self, 'export_frame_range') col.prop(self, 'export_frame_step') col.prop(self, 'export_move_keyframes') col.prop(self, 'export_force_sampling') else: col.prop(self, 'export_current_frame') col.prop(self, 'export_skins') if self.export_skins: col.prop(self, 'export_bake_skins') col.prop(self, 'export_morph') if self.export_morph: col.prop(self, 'export_morph_normal') if self.export_morph_normal: col.prop(self, 'export_morph_tangent') addon_prefs = context.user_preferences.addons[__name__].preferences if addon_prefs.experimental: col = layout.box().column() col.label('Experimental:', icon='RADIO') col.prop(self, 'export_lights') col.prop(self, 'export_displacement') row = layout.row() row.operator( GLTF2ExportSettings.bl_idname, GLTF2ExportSettings.bl_label, icon="%s" % "PINNED" if self.will_save_settings else "UNPINNED") class ExportGLTF2_GLTF(bpy.types.Operator, ExportGLTF2_Base, ExportHelper): '''Export scene as glTF 2.0 file''' bl_idname = 'export_scene.gltf' bl_label = 'Export glTF 2.0' filename_ext = '.gltf' filter_glob = StringProperty(default='*.gltf', options={'HIDDEN'}) export_format = 'ASCII' class ExportGLTF2_GLB(bpy.types.Operator, ExportGLTF2_Base, ExportHelper): '''Export scene as glTF 2.0 file''' bl_idname = 'export_scene.glb' bl_label = 'Export glTF 2.0 binary' filename_ext = '.glb' filter_glob = StringProperty(default='*.glb', options={'HIDDEN'}) export_format = 'BINARY' def menu_func_export_gltf(self, context): self.layout.operator(ExportGLTF2_GLTF.bl_idname, text='glTF 2.0 (.gltf)') def menu_func_export_glb(self, context): self.layout.operator(ExportGLTF2_GLB.bl_idname, text='glTF 2.0 (.glb)') from bpy.types import AddonPreferences class ExportGLTF2_AddonPreferences(AddonPreferences): bl_idname = __name__ experimental = BoolProperty(name='Enable experimental glTF export settings', default=False) def draw(self, context): layout = self.layout layout.prop(self, "experimental") class ImportglTF2(Operator, ImportHelper): bl_idname = 'import_scene.gltf' bl_label = "glTF 2.0 (.gltf/.glb)" filename_ext = ".gltf" filter_glob = StringProperty(default="*.gltf;*.glb", options={'HIDDEN'}) loglevel = bpy.props.EnumProperty(items=Log.getLevels(), description="Log Level", default=Log.default()) def execute(self, context): return self.import_gltf2(context) def import_gltf2(self, context): bpy.context.scene.render.engine = 'CYCLES' self.gltf = glTFImporter(self.filepath, self.loglevel) self.gltf.log.critical("Starting loading glTF file") success, txt = self.gltf.read() if not success: self.report({'ERROR'}, txt) return {'CANCELLED'} self.gltf.log.critical("Data are loaded, start creating Blender stuff") self.gltf.blender_create() self.gltf.debug_missing() self.gltf.log.critical("glTF import is now finished") self.gltf.log.removeHandler(self.gltf.log_handler) # Switch to newly created main scene bpy.context.screen.scene = bpy.data.scenes[self.gltf.blender.scene] return {'FINISHED'} def menu_func_import(self, context): self.layout.operator(ImportglTF2.bl_idname, text=ImportglTF2.bl_label) def register(): bpy.utils.register_module(__name__) bpy.types.INFO_MT_file_export.append(menu_func_export_gltf) bpy.types.INFO_MT_file_export.append(menu_func_export_glb) bpy.types.INFO_MT_file_import.append(menu_func_import) def unregister(): bpy.utils.unregister_module(__name__) bpy.types.INFO_MT_file_export.remove(menu_func_export_gltf) bpy.types.INFO_MT_file_export.remove(menu_func_export_glb) bpy.types.INFO_MT_file_import.remove(menu_func_import)
[ "nopper@ux3d.io" ]
nopper@ux3d.io
8b09a98c3ac1acf69e5c84f6bbeeb54671c20bc6
11ce41733d6f31153fe14f800c9dd0be18615862
/news/admin.py
50285420a545e93e7a3d322e73e11bb5a4d627f4
[ "MIT" ]
permissive
techacademypython/django_image_crop_views
6ff6731944f5d09721452a71b0745089d1b035ef
2f9c51ae80705dc23607e157baa4f5767957a2f1
refs/heads/master
2023-05-05T13:12:23.642970
2019-09-03T16:38:24
2019-09-03T16:38:24
206,105,932
0
0
MIT
2022-11-22T04:13:41
2019-09-03T15:07:05
Python
UTF-8
Python
false
false
375
py
from django.contrib import admin from image_cropping import ImageCroppingMixin # Register your models here. from news.models import NewsModel class NewsModelAdmin(ImageCroppingMixin, admin.ModelAdmin): readonly_fields = ["preview_count"] fields = [ "image", "name", "text", "cropping", "preview_count" ] admin.site.register(NewsModel, NewsModelAdmin)
[ "munisisazade@gmail.com" ]
munisisazade@gmail.com
b897b084b288350d1a287661007953393d395943
0fccee4c738449f5e0a8f52ea5acabf51db0e910
/genfragments/EightTeV/BprimeBprime/BprimeBprimeToBHBHinc_M_800_TuneZ2star_8TeV_madgraph_cff.py
352c4947ebcf1ce31ccf35f0dd2e24c3165cb26a
[]
no_license
cms-sw/genproductions
f308ffaf3586c19b29853db40e6d662e937940ff
dd3d3a3826343d4f75ec36b4662b6e9ff1f270f4
refs/heads/master
2023-08-30T17:26:02.581596
2023-08-29T14:53:43
2023-08-29T14:53:43
11,424,867
69
987
null
2023-09-14T12:41:28
2013-07-15T14:18:33
Python
UTF-8
Python
false
false
4,231
py
import FWCore.ParameterSet.Config as cms #from Configuration.Generator.PythiaUEZ2Settings_cfi import * from Configuration.Generator.PythiaUEZ2starSettings_cfi import * generator = cms.EDFilter("Pythia6HadronizerFilter", pythiaHepMCVerbosity = cms.untracked.bool(False), maxEventsToPrint = cms.untracked.int32(0), pythiaPylistVerbosity = cms.untracked.int32(0), comEnergy = cms.double(8000.0), PythiaParameters = cms.PSet( pythiaUESettingsBlock, processParameters = cms.vstring( 'PMAS(25,1)=125.00D0 !mass of Higgs', 'MSTP(1) = 4', 'MSEL=7 ! User defined processes', 'MWID(7)=2', 'MSTJ(1)=1 ! Fragmentation/hadronization on or off', 'MSTP(61)=1 ! Parton showering on or off', 'PMAS(5,1)=4.8 ! b quark mass', #from Spring11 4000040 'PMAS(6,1)=172.5 ! t quark mass', #from Spring11 4000040 'PMAS(7,1) = 800.0D0 ! bprime quarks mass', 'PMAS(7,2) = 8.000D0 ! bprime quark width', 'PMAS(7,3) = 80.00D0 ! Max value above which the BW shape is truncated', 'VCKM(1,1) = 0.97414000D0', 'VCKM(1,2) = 0.22450000D0', 'VCKM(1,3) = 0.00420000D0', 'VCKM(1,4) = 0.02500000D0', 'VCKM(2,1) = 0.22560000D0', 'VCKM(2,2) = 0.97170000D0', 'VCKM(2,3) = 0.04109000D0', 'VCKM(2,4) = 0.05700000D0', 'VCKM(3,1) = 0.00100000D0', 'VCKM(3,2) = 0.06200000D0', 'VCKM(3,3) = 0.91000000D0', 'VCKM(3,4) = 0.41000000D0', 'VCKM(4,1) = 0.01300000D0', 'VCKM(4,2) = 0.04000000D0', 'VCKM(4,3) = 0.41000000D0', 'VCKM(4,4) = 0.91000000D0', 'MDME(56,1)=0 ! g b4', 'MDME(57,1)=0 ! gamma b4', 'MDME(58,1)=0 ! Z0 b', 'MDME(59,1)=0 ! W u', 'MDME(60,1)=0 ! W c', 'MDME(61,1)=0 ! W t', 'MDME(62,1)=0 ! W t4', 'KFDP(63,2)=5 ! defines H0 b', 'MDME(63,1)=1 ! h0 b4', 'MDME(64,1)=-1 ! H- c', 'MDME(65,1)=-1 ! H- t', 'BRAT(56) = 0.0D0', 'BRAT(57) = 0.0D0', 'BRAT(58) = 0.0D0', 'BRAT(59) = 0.0D0', 'BRAT(60) = 0.0D0', 'BRAT(61) = 0.0D0', 'BRAT(62) = 0.0D0', 'BRAT(63) = 1.0D0', 'BRAT(64) = 0.0D0', 'BRAT(65) = 0.0D0', 'MDME(210,1)=1 !Higgs decay into dd', 'MDME(211,1)=1 !Higgs decay into uu', 'MDME(212,1)=1 !Higgs decay into ss', 'MDME(213,1)=1 !Higgs decay into cc', 'MDME(214,1)=1 !Higgs decay into bb', 'MDME(215,1)=1 !Higgs decay into tt', 'MDME(216,1)=1 !Higgs decay into', 'MDME(217,1)=1 !Higgs decay into Higgs decay', 'MDME(218,1)=1 !Higgs decay into e nu e', 'MDME(219,1)=1 !Higgs decay into mu nu mu', 'MDME(220,1)=1 !Higgs decay into tau nu tau', 'MDME(221,1)=1 !Higgs decay into Higgs decay', 'MDME(222,1)=1 !Higgs decay into g g', 'MDME(223,1)=1 !Higgs decay into gam gam', 'MDME(224,1)=1 !Higgs decay into gam Z', 'MDME(225,1)=1 !Higgs decay into Z Z', 'MDME(226,1)=1 !Higgs decay into W W', ), # This is a vector of ParameterSet names to be read, in this order parameterSets = cms.vstring('pythiaUESettings', 'processParameters') ), jetMatching = cms.untracked.PSet( scheme = cms.string("Madgraph"), mode = cms.string("auto"), # soup, or "inclusive" / "exclusive" MEMAIN_etaclmax = cms.double(5.0), MEMAIN_qcut = cms.double(-1), MEMAIN_nqmatch = cms.int32(-1), MEMAIN_minjets = cms.int32(-1), MEMAIN_maxjets = cms.int32(-1), MEMAIN_showerkt = cms.double(0), MEMAIN_excres = cms.string(''), outTree_flag = cms.int32(0) ) ) ProductionFilterSequence = cms.Sequence(generator)
[ "sha1-5c9a4926c1ea08b633689ec734e2440da58b8c56@cern.ch" ]
sha1-5c9a4926c1ea08b633689ec734e2440da58b8c56@cern.ch
0df124adf5c65ca34ee6bc7af7252e7d3ae24fad
0b0570c7799b5997bd7ed828d52142de591229b8
/fizzbuzz.py
178fcfb5fd13c1d0e2a9d0257365682dc67f1543
[]
no_license
triaddojo/fizzbuzz-jiro
2c8c7d56fe702cf023010d45d5d7c897f8266874
c643d4ea0c861aa49e61f554398597272a1cd572
refs/heads/master
2020-05-26T09:05:34.137623
2019-05-23T07:26:45
2019-05-23T07:26:45
188,178,480
0
2
null
2019-05-23T07:52:48
2019-05-23T06:55:20
Python
UTF-8
Python
false
false
208
py
for number in range(1, 101): if number % 15 == 0: print("fizzbuzz") elif number % 3 == 0: print("fizz") elif number % 5 == 0: print("buzz") else: print(number)
[ "akira.wakatsuki@tri-ad.global" ]
akira.wakatsuki@tri-ad.global
fa0707a68b09bbfbbd1c0b2ee3e6266f543f5dc2
f073c033419dffa5237d470c0206f97adaf9db99
/database_learning/employee.py
b13488a97ac3565a636dbfba264664844f9e15ca
[]
no_license
schnippo/spanishconjugator
618b858580ca46c1a2fee7aae167025e3e72de41
ab8477ef94bf077b339d046b7e82a923be90ca10
refs/heads/master
2020-04-16T14:04:28.909452
2019-01-14T12:07:00
2019-01-14T12:07:00
165,653,824
0
0
null
null
null
null
UTF-8
Python
false
false
461
py
class Employee: def __init__(self, first, last, pay): self.first = first self.last = last self.pay = pay @property def email(self): return '{}.{}@company.com'.format(self.first, self.last) @property def fullname(self): return '{} {}'.format(self.first, self.last) def __repr__(self): return "Employee('{}','{}','{}'".format(self.first,self.last, self.pay) # john = Employee('John', 'Buttle', 50000) # print(john.email) # print(repr(john))
[ "noreply@github.com" ]
noreply@github.com
a4e0c192f3c8f4463eae05876b00114d00ab91c7
8ce23f191870868c86c7616882e6043b1102cb0d
/tools/text_processing/join_files_on_column_fuzzy/join_files_on_column_fuzzy.py
1e19f1dcfe4a4d3ab0743078894f5c196b0b2559
[]
no_license
StevenVerbruggen/galaxytools
56f99d0d629cb6d9e3db290c64f30b920de04f26
7d7365197e2cba2eb048121c9f0ee5546f06c520
refs/heads/master
2021-01-16T17:51:39.721403
2020-12-01T08:35:51
2020-12-01T08:35:51
100,017,016
0
0
null
2017-08-11T09:42:20
2017-08-11T09:42:20
null
UTF-8
Python
false
false
4,755
py
#!/usr/bin/env python import os import argparse import sys def main(args): if args.header: h1 = True h2 = True else: h1 = False h2 = False cache = list() out = open(args.outfile, 'w+') write_buffer = list() def _readline(header = False): with open(args.f2) as handle2: for line in handle2: line = line.strip() if header: header = False yield line continue if not line: continue columns = line.split(args.sep) value2 = columns[args.c2-1] yield columns, float(value2) def fill_cache(): try: cache.append(next(it)) except StopIteration: pass it = _readline(header = h2) with open(args.f1) as handle1: for line in handle1: line = line.strip() if h1: h1 = False seconda_header = next(it) if args.add_distance: out.write('%s\t%s\t%s\n' % (line, seconda_header, args.unit)) else: out.write('%s\t%s\n' % (line, seconda_header)) continue if not line: continue columns = line.split(args.sep) value1 = float(columns[args.c1-1]) _cache = list() fill_cache() while cache: _c, value2 = cache.pop(0) upper_bound = value1 + args.distance if args.unit == 'absolute': if value2 <= upper_bound and value2 >= (value1 - args.distance): line_template = '%s\n' abs_dist = abs(value1 - value2) if args.add_distance: line_template = '%s\t' + str(abs_dist) + '\n' write_buffer.append([abs_dist, line_template % '\t'.join( columns + _c )]) _cache.append([_c, value2]) fill_cache() elif value2 > upper_bound: # if the value from list 2 is bigger then the current value, he will be taken into the next round _cache.append([_c, value2]) elif value2 < upper_bound: # if the value from list 2 is smaller then the currecnt value, check the next one of list 2 fill_cache() elif args.unit == 'ppm': ppm_dist = abs((value1 - value2) / value1 * 1000000) if ppm_dist <= args.distance: line_template = '%s\n' if args.add_distance: line_template = '%s\t' + str(ppm_dist) + '\n' write_buffer.append([ppm_dist, line_template % '\t'.join( columns + _c )]) _cache.append([_c, value2]) fill_cache() elif ppm_dist > args.distance: _cache.append([_c, value2]) elif ppm_dist < args.distance: fill_cache() if args.closest and write_buffer: write_buffer.sort(key=lambda x: x[0]) out.write(write_buffer[0][1]) else: for _dist, line in write_buffer: out.write(line) write_buffer = list() cache = _cache out.close() if __name__ == '__main__': parser = argparse.ArgumentParser(description='Merge two files on a common column the fuzzy way.') parser.add_argument('--f1', required=True) parser.add_argument('--f2', required=True) parser.add_argument('--c1', type=int, required=True, help="Column in file 1 to be merged on.") parser.add_argument('--c2', type=int, required=True, help="Column in file 2 to be merged on.") parser.add_argument('--outfile', required=True) parser.add_argument('--header', action='store_true', help="The files have a header line at the beginning.") parser.add_argument('--closest', action='store_true', help="Only report the closest match.") parser.add_argument('--add_distance', action='store_true', help="Add addional column with the distance between the two values.") parser.add_argument('--sep', type=str, default="\t", help="Files are separated by this separator.") parser.add_argument('--distance', type=float, default="0.2", help="Maximal allowed distance.") parser.add_argument('--unit', choices=['ppm', 'absolute'], default='absolute') args = parser.parse_args() main(args)
[ "bjoern.gruening@gmail.com" ]
bjoern.gruening@gmail.com
056dc9af3862ebbbd133e049bcacd156c6262d0c
18d53974bd1eeaa22fc93715041b11903f73a434
/SEIRMU/new/OutofSample/draw_picture/paint_line.py
1ffef5d7f8fcad2403c7a6d5241c4b397f23155a
[ "MIT" ]
permissive
KL-ice/SEIR-AIM
b363451ea1f7d1c099ad2fe29fbb4657ce864b1a
c12a67186a6d8a59deb56b7a29ce86f170fc9d0c
refs/heads/main
2023-07-18T23:22:53.107921
2021-09-30T04:55:54
2021-09-30T04:55:54
411,366,642
0
0
null
null
null
null
UTF-8
Python
false
false
20,652
py
import numpy as np import pandas as pd import datetime as dt import matplotlib.pyplot as plt import matplotlib.dates as mdates import matplotlib.ticker as ticker from matplotlib.font_manager import * from pylab import * from matplotlib.pyplot import MultipleLocator from matplotlib.ticker import AutoMinorLocator from read_data import * from config import * import matplotlib.font_manager def paint_confirm(pred_data, truth_data): lenth_pred = len(pred_data) lenth_truth = len(truth_data) x_pred = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data x_truth = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_truth] y_truth = truth_data #调整图片大小 plt.figure(figsize=(8,6)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'in' matplotlib.rcParams['ytick.direction'] = 'in' tick_params(which='both',top='on',bottom='on',left='on',right='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(6000000,130000000) ax.set_ylim(6000000,60000000) # 设置网格线 ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_linewidth(1) ax.spines['left'].set_linewidth(1) ax.spines['right'].set_linewidth(1) plt.plot(x_pred,y_pred,label="Prediction Confirmed Cases",color="#D95319",linewidth=3) plt.plot(x_truth,y_truth,label="Confirmed Cases by Database",color="#0072BD",linewidth=3) plt.tick_params(labelsize=15) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=18) plt.ylabel("Confirmed Cases",fontsize=18) # plt.title("Confirmed Cases",fontsize=20) # plt.gcf().autofmt_xdate() # ax.get_yaxis().get_major_formatter().set_scientific(False) plt.legend(loc='center left', bbox_to_anchor=(0.03, 0.78), ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.2, fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.savefig('./result/confirm.jpg', dpi=1200, pad_inches=0.0) def paint_unemployment(pred_data, truth_data): lenth_pred = len(pred_data) lenth_truth = len(truth_data) x_pred = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data x_truth = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_truth] y_truth = truth_data #调整图片大小 plt.figure(figsize=(8,6)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'in' matplotlib.rcParams['ytick.direction'] = 'in' tick_params(which='both',top='on',bottom='on',left='on',right='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) # 设置网格线 ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_linewidth(1) ax.spines['left'].set_linewidth(1) ax.spines['right'].set_linewidth(1) plt.plot(x_pred,y_pred,label="Prediction Unemployment Rate",color="#D95319",linewidth=3) plt.plot(x_truth,y_truth,label="Unemployment Rate by Database",color="#0072BD",linewidth=3) plt.tick_params(labelsize=15) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=18) plt.ylabel("Unemployment Rate",fontsize=18) plt.title("Unemployment Rate",fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) plt.legend(loc='center right',bbox_to_anchor=(1, 0.7), ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.savefig('./result/unemployment.svg', dpi=1200, pad_inches=0.0) def paint_inf(pred_data, truth_data): lenth_pred = len(pred_data) lenth_truth = len(truth_data) x_pred = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data x_truth = [dt.datetime.strptime('2020-09-26', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_truth] y_truth = truth_data #调整图片大小 plt.figure(figsize=(8,6)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'in' matplotlib.rcParams['ytick.direction'] = 'in' tick_params(which='both',top='on',bottom='on',left='on',right='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) # 设置网格线 ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_linewidth(1) ax.spines['left'].set_linewidth(1) ax.spines['right'].set_linewidth(1) plt.plot(x_pred,y_pred,label="Prediction Infection Rate",color="#D95319",linewidth=3) plt.plot(x_truth,y_truth,label="Infection Rate by Database",color="#0072BD",linewidth=3) plt.tick_params(labelsize=15) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=18) plt.ylabel("Infection Rate",fontsize=18) plt.title("Infection Rate",fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) plt.legend(loc='center right',bbox_to_anchor=(1, 0.7), ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.subplots_adjust(left=0.15, top=None, wspace=None, hspace=None) plt.savefig('./result/infection_rate.svg', dpi=1200, pad_inches=0.0) def paint_Rt(pred_data, truth_data): lenth_pred = len(pred_data) lenth_truth = len(truth_data) x_pred = [dt.datetime.strptime('2020-03-20', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data x_truth = [dt.datetime.strptime('2020-03-20', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_truth] y_truth = truth_data #调整图片大小 plt.figure(figsize=(8,6)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'in' matplotlib.rcParams['ytick.direction'] = 'in' tick_params(which='both',top='on',bottom='on',left='on',right='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) # 设置网格线 ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_linewidth(1) ax.spines['left'].set_linewidth(1) ax.spines['right'].set_linewidth(1) plt.plot(x_pred,y_pred,label="Prediction Rt",color="#D95319",linewidth=3) plt.plot(x_truth,y_truth,label="Rt by Database",color="#0072BD",linewidth=3) plt.tick_params(labelsize=15) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=18) plt.ylabel("Rt",fontsize=18) plt.title("Reproduction Number", fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) plt.legend(loc='center right',bbox_to_anchor=(0.98, 0.73),ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() def paint_blm_confirm(pred_data): lenth_pred = len(pred_data) x_pred = [dt.datetime.strptime('2020-06-27', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data #调整图片大小 plt.figure(figsize=(12,5.5)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'out' matplotlib.rcParams['ytick.direction'] = 'out' tick_params(which='both',bottom='on',left='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) # 设置网格线 # ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) # ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_color('none') ax.spines['left'].set_linewidth(1) # ax.spines['right'].set_linewidth(1) ax.spines['right'].set_color('none') plt.plot(x_pred,y_pred,label="prediction Rt",color="#D95319",linewidth=3) plt.tick_params(labelsize=18) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=23, fontweight='bold') plt.ylabel("Increased confirmed \ncases caused by BLM",fontsize=23, fontweight='bold') # plt.title("Increased Cumulative Confirmed Cases Caused by BLM", fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) # plt.legend(loc='upper right',ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, # fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.savefig('./result/blm confirm.svg', dpi=1200, pad_inches=0.05, bbox_inches = 'tight') def paint_blm_unemployment(pred_data): lenth_pred = len(pred_data) x_pred = [dt.datetime.strptime('2020-06-27', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data #调整图片大小 plt.figure(figsize=(12,5.5)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'out' matplotlib.rcParams['ytick.direction'] = 'out' tick_params(which='both',bottom='on',left='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) # 设置网格线 # ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) # ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_color('none') ax.spines['left'].set_linewidth(1) # ax.spines['right'].set_linewidth(1) ax.spines['right'].set_color('none') plt.plot(x_pred,y_pred,label="prediction Rt",color="#D95319",linewidth=3) plt.tick_params(labelsize=18) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=23, fontweight='bold') plt.ylabel("Increased unemployment \nrate caused by BLM",fontsize=23, fontweight='bold') # plt.title("Increased Unemployment Rate Caused by BLM", fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) # plt.legend(loc='upper right',ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, # fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.savefig('./result/blm unemployment.svg', dpi=1200, pad_inches=0.05, bbox_inches = 'tight') def paint_blm_index(pred_data): lenth_pred = len(pred_data) x_pred = [dt.datetime.strptime('2020-05-20', '%Y-%m-%d').date() + dt.timedelta(days = x) for x in range(1000)][:lenth_pred] y_pred = pred_data #调整图片大小 plt.figure(figsize=(12,5.5)) # 设定坐标轴刻度朝内 matplotlib.rcParams['xtick.direction'] = 'out' matplotlib.rcParams['ytick.direction'] = 'out' tick_params(which='both',bottom='on',left='on') #设定字体 plt.rc('font',family='Times New Roman') #设定标题字体 plt.rcParams['font.sans-serif'] = ['Times New Roman'] ax = plt.gca() # ax.set_ylim(0,13000000) ax.set_ylim(0,1) # 设置网格线 # ax.grid(axis='x',which='major',color= 'gray',alpha = 0.4) # ax.grid(axis='y',which='major',color= 'gray',alpha = 0.4) # 设置坐标轴边框宽度 # ax.spines['bottom'].set_color('#0072BD') ax.spines['bottom'].set_linewidth(1) ax.spines['top'].set_color('none') ax.spines['left'].set_linewidth(1) # ax.spines['right'].set_linewidth(1) ax.spines['right'].set_color('none') plt.plot(x_pred,y_pred,label="prediction Rt",color="#0072BD",linewidth=3, clip_on=False, zorder=3) plt.tick_params(labelsize=18) labels = ax.get_xticklabels() + ax.get_yticklabels() [label.set_fontname('Times New Roman') for label in labels] plt.xlabel("Date",fontsize=23, fontweight='bold') plt.ylabel("\nBLM Index",fontsize=23, fontweight='bold') # plt.title("BLM Index", fontsize=20) plt.gcf().autofmt_xdate() ax.get_yaxis().get_major_formatter().set_scientific(False) # plt.legend(loc='upper right',ncol=1,borderaxespad=0.3, edgecolor='gray',labelspacing=0.1,handletextpad=0.7,columnspacing=0.3,handlelength=1.7,borderpad=0.1, # fancybox=True,frameon=True,fontsize=20,shadow=False, framealpha=0.5) # plt.show() plt.subplots_adjust(left=0.1, bottom=0.25, right=None, top=None) plt.savefig('./result/blm index.svg', dpi=1200, pad_inches=0.05, bbox_inches = 'tight') if __name__ == "__main__": # paint confirm cases # pred_confirm = read_pred_confirm(pred_confirm_path) # truth_confirm = read_truth_confirm(truth_confirm_path) # paint_confirm(pred_confirm, truth_confirm) # paint unemployment rate # pred_unemployment = read_pred_unemployment(pred_unemployment_path) # truth_unemployment = read_truth_unemployment(truth_unemployment_path) # paint_unemployment(pred_unemployment, truth_unemployment) # paint unemployment rate # pred_Rt = read_pred_Rt(pred_Rt_path) # truth_Rt = read_truth_Rt(truth_Rt_path) # paint_Rt(pred_Rt, truth_Rt) # paint blm different confirm pred_blm_confirm = read_blm_confirm(blm_confirm_path, noblm_pred_confirm_path) print(pred_blm_confirm) paint_blm_confirm(pred_blm_confirm) # paint blm different unemployment pred_blm_unemployment = read_blm_unemployment(blm_unemployment_path, noblm_pred_unemployment_path) print(pred_blm_unemployment) paint_blm_unemployment(pred_blm_unemployment) blm_index = read_blm_inedx(blm_index_path) paint_blm_index(blm_index) # pred_inf = [0.00954041164368391, 0.010497291572391987, 0.011120161972939968, 0.011480739340186119, 0.011714714579284191, 0.011915579438209534, 0.01210733037441969, 0.012298648245632648, 0.012515188194811344, 0.012772245332598686, 0.013041410595178604, 0.013296606950461864, 0.013511604629456997, 0.013682755641639233, 0.013837314210832119, 0.013980587013065815, 0.01415819302201271, 0.014362465590238571, 0.014596445485949516, 0.014872822910547256, 0.015195727348327637, 0.015555099584162235, 0.015946270897984505, 0.016329947859048843, 0.016726627945899963, 0.017150476574897766, 0.017611948773264885, 0.01811842992901802, 0.0186744574457407, 0.019233373925089836, 0.019821060821413994, 0.02044534869492054, 0.021105162799358368, 0.0217897966504097, 0.022479863837361336, 0.023150412365794182, 0.02383909747004509, 0.024510089308023453, 0.025131208822131157, 0.025675132870674133, 0.026120556518435478, 0.026453059166669846, 0.026665957644581795, 0.026752926409244537, 0.026725631207227707, 0.026600774377584457, 0.02639753557741642, 0.02613612823188305, 0.025836963206529617, 0.02552018314599991, 0.025211604312062263, 0.024926593527197838, 0.0246753953397274, 0.024463104084134102, 0.02429005317389965, 0.02415289357304573, 0.024046216160058975, 0.02396765537559986, 0.023909378796815872, 0.02386578358709812, 0.023832842707633972, 0.02380770817399025, 0.023788345977663994, 0.023773277178406715, 0.023760177195072174, 0.02374868467450142, 0.023738589137792587, 0.023729639127850533, 0.02372164838016033, 0.023714477196335793, 0.023708004504442215, 0.023702135309576988, 0.023696795105934143, 0.023691847920417786, 0.023687273263931274, 0.023683026432991028, 0.023679085075855255, 0.02367541566491127, 0.023671990260481834, 0.023668792098760605, 0.023665815591812134, 0.02366304025053978, 0.0236604493111372, 0.02365802973508835, 0.023655762895941734, 0.023653637617826462, 0.023651644587516785, 0.023649776354432106, 0.02364801988005638, 0.023646365851163864, 0.02364480309188366, 0.023643335327506065, 0.02364194579422474, 0.02364063635468483, 0.02363939955830574, 0.023638226091861725, 0.023637115955352783, 0.023636065423488617, 0.023635070770978928, 0.02363412454724312, 0.02363322302699089, 0.023632371798157692, 0.023631559684872627, 0.023630788549780846, 0.02363005466759205, 0.023629354313015938, 0.023628689348697662, 0.023628052324056625, 0.023627446964383125, 0.023626862093806267, 0.023626312613487244, 0.02362578734755516, 0.023625284433364868, 0.02362479828298092, 0.02362433634698391, 0.02362389862537384, 0.023623475804924965, 0.023623069748282433, 0.023622682318091393, 0.023622315376996994, 0.023621955886483192, 0.023621613159775734, 0.023621289059519768, 0.023620974272489548, 0.023620672523975372, 0.02362038940191269, 0.023620113730430603, 0.023619843646883965, 0.02361958660185337, 0.023619340732693672, 0.02361910417675972, 0.023618875071406364, 0.023618659004569054, 0.02361844852566719, 0.023618245497345924, 0.023618051782250404, 0.02361786924302578, 0.023617688566446304, 0.023617513477802277, 0.023617349565029144, 0.02361719124019146, 0.023617036640644073, 0.023616887629032135, 0.023616742342710495, 0.0236166063696146, 0.023616474121809006] # truth_inf = [0.010669372654577562, 0.010743240777903455, 0.011088392290809719, 0.011265271501965275, 0.011462000790084845, 0.011562583441632683, 0.01156611761364354, 0.011653536320369123, 0.012039351765481387, 0.012346229607414416, 0.012462875007794968, 0.012730140084263297, 0.012777295531378526, 0.012847473469267073, 0.013122772616106913, 0.01342940903331596, 0.013825499055621235, 0.014240617405671027, 0.014371340671943785, 0.014751034383418807, 0.014813940578132718, 0.01513148986078562, 0.01564035684584562, 0.015997450537550747, 0.016503159356639333, 0.016873487000925773, 0.01725682576936934, 0.017514821092231986, 0.01788690190946341, 0.01866813324533322, 0.019538106392953718, 0.020196194105407496, 0.02105420048583124, 0.02160278622977868, 0.022103806515040448, 0.02257332422559926, 0.02320140776025004, 0.02374017975137745, 0.024515457371323938, 0.0247350483605454, 0.025252417470816486, 0.02557000287228027, 0.025888138607007626, 0.026453451187721327, 0.027004958855033203, 0.027219782122427445, 0.027054776460013908, 0.026819453020947334, 0.026500476140760496, 0.026123419498499372, 0.025989503849476524, 0.025796745229456727, 0.02558611066731834, 0.025133127836314745, 0.024320086338151945, 0.023748778904985637, 0.0231924095516537] # paint_inf(pred_inf, truth_inf)
[ "wbkingice@sina.com" ]
wbkingice@sina.com
41ef33c1c1af378a664ea82f485c5a12ebeedd1c
a0fb29f99a852089193e4cc9a11e7263dc3f8b5f
/mayan/apps/metadata/literals.py
aba1309e370f89d0f6259a24ca393df9dc3e1f1c
[ "Apache-2.0" ]
permissive
ikang9712/Mayan-EDMS
0e22a944d63657cea59c78023b604a01a622b52a
d6e57e27a89805329fe0c5582caa8e17882d94e6
refs/heads/master
2023-07-28T19:41:55.269513
2021-09-07T14:16:14
2021-09-07T14:16:14
402,884,683
1
0
NOASSERTION
2021-09-03T20:00:09
2021-09-03T20:00:09
null
UTF-8
Python
false
false
227
py
from .parsers import MetadataParser from .validators import MetadataValidator DEFAULT_METADATA_AVAILABLE_VALIDATORS = MetadataValidator.get_import_paths() DEFAULT_METADATA_AVAILABLE_PARSERS = MetadataParser.get_import_paths()
[ "roberto.rosario@mayan-edms.com" ]
roberto.rosario@mayan-edms.com
2eebba53d96810f87e30cf377556f367d5ae17b1
cc91403e4302d70127562591ab3fda7a212e6312
/asqcenv/lib/python3.9/site-packages/asqc/asqc.py
ade74103aca22e8cf2e1c774024a023bed7f5a03
[]
no_license
gklyne/asqc
fdfb59e243f92968e2147ef28ce6c85394d9cab0
1fcd4d80727b8385e8707be12d1e45fe26d2229a
refs/heads/master
2022-07-28T09:05:07.923525
2022-07-14T19:44:52
2022-07-14T19:44:52
3,959,374
3
1
null
null
null
null
UTF-8
Python
false
false
26,669
py
#!/usr/bin/env python """ ASQC - A SPARQL query client """ import sys import os import os.path import urllib.parse import urllib import urllib.request # import io.StringIO import io import json import re import optparse import logging import traceback from .SparqlHttpClient import SparqlHttpClient from .SparqlXmlResults import writeResultsXML from .StdoutContext import SwitchStdout from .StdinContext import SwitchStdin import rdflib # Set up to use SPARQL # rdflib.plugin.register( # 'sparql', rdflib.query.Processor, # 'rdfextras.sparql.processor', 'Processor') # rdflib.plugin.register( # 'sparql', rdflib.query.Result, # 'rdfextras.sparql.query', 'SPARQLQueryResult') # Register serializers (needed?) #rdflib.plugin.register('n3', Serializer, # 'rdflib.plugins.serializers.n3','N3Serializer') #rdflib.plugin.register('turtle', Serializer, # 'rdflib.plugins.serializers.turtle', 'TurtleSerializer') #rdflib.plugin.register('nt', Serializer, # 'rdflib.plugins.serializers.nt', 'NTSerializer') #rdflib.plugin.register('xml', Serializer, # 'rdflib.plugins.serializers.rdfxml', 'XMLSerializer') #rdflib.plugin.register('pretty-xml', Serializer, # 'rdflib.plugins.serializers.rdfxml', 'PrettyXMLSerializer') #rdflib.plugin.register('json-ld', Serializer, # 'rdflib.plugins.serializers.rdfxml', 'XMLSerializer') #plugin.register('json-ld', Serializer, # 'rdfextras.serializers.jsonld', 'JsonLDSerializer') # Type codes and mapping for RDF and query variable p[arsing and serializing RDFTYP = ["RDFXML","N3","TURTLE","NT","JSONLD","RDFA","HTML5"] VARTYP = ["JSON","CSV","XML"] RDFTYPPARSERMAP = ( { "RDFXML": "xml" , "N3": "n3" , "TURTLE": "n3" , "NT": "nt" , "JSONLD": "jsonld" , "RDFA": "rdfa" , "HTML5": "rdfa+html" }) RDFTYPSERIALIZERMAP = ( { "RDFXML": "pretty-xml" , "N3": "n3" , "TURTLE": "turtle" , "NT": "nt" , "JSONLD": "jsonld" }) # Logging object log = logging.getLogger(__name__) from . import __init__ class asqc_settings(object): VERSION = "1.0.9" # __init__.__version__ @@@@ # Helper function for templated SPARQL results formatting and parsing def formatBindings(template, bindings): """ Return bindings formatted with supplied template """ formatdict = {} for (var, val) in bindings.iteritems(): formatdict[var] = val["value"] if val["type"] == "bnode": vf = "_:%(value)s" elif val["type"] == "uri": vf = "<%(value)s>" elif val["type"] == "literal": vf = '"%(value)s"' elif val["type"] == "typed-literal": vf = '"%(value)s"^^<%(datatype)s>' formatdict[var+"_repr"] = vf%val return template.decode(encoding='string_escape')%formatdict # Helper function for CSV formatting query result from JSON def char_escape(c): if c == '"': return '""' if ord(c) >= 128: return r"\u" + "%04x"%ord(c) return c def termToCSV(result): if result == None: return None resval = result['value'] restyp = result['type'] if restyp == "uri": return "<" + resval + ">" if restyp == "bnode": return "_:" + resval # strval = '"' + resval.replace('"', '""') + '"' strval = '"' + "".join([char_escape(c) for c in resval]) + '"' strlang = result.get('xml:lang', None) if restyp == "literal": if strlang: return strval + '@' + strlang else: return strval if restyp == "typed-literal": return strval + '^^' + result['datatype'] raise rdflib.query.ResultException('Unknown term type: %s (%s)'%(term, type(term))) # Helper functions for JSON formatting and parsing # Mostly copied from rdflib SPARQL code (rdfextras/sparql/results/jsonresults) def termToJSON(term): if isinstance(term, rdflib.URIRef): return { 'type': 'uri', 'value': str(term) } elif isinstance(term, rdflib.Literal): if term.datatype!=None: return { 'type': 'typed-literal', 'value': str(term), 'datatype': str(term.datatype) } else: r={'type': 'literal', 'value': str(term) } if term.language!=None: r['xml:lang']=term.language return r elif isinstance(term, rdflib.BNode): return { 'type': 'bnode', 'value': str(term) } elif term==None: return None else: raise rdflib.query.ResultException('Unknown term type: %s (%s)'%(term, type(term))) def bindingToJSON(binding): res={} for var in binding: t = termToJSON(binding[var]) if t != None: res[str(var)] = t return res def parseJsonTerm(d): """rdflib object (Literal, URIRef, BNode) for the given json-format dict. input is like: { 'type': 'uri', 'value': 'http://famegame.com/2006/01/username' } { 'type': 'bnode', 'value': '123abc456' } { 'type': 'literal', 'value': 'drewp' } { 'type': 'literal', 'value': 'drewp', xml:lang="en" } { 'type': 'typed-literal', 'value': '123', datatype="http://(xsd)#int" } """ t = d['type'] if t == 'uri': return rdflib.URIRef(d['value']) elif t == 'literal': if 'xml:lang' in d: return rdflib.Literal(d['value'], lang=d['xml:lang']) return rdflib.Literal(d['value']) elif t == 'typed-literal': return rdflib.Literal(d['value'], datatype=rdflib.URIRef(d['datatype'])) elif t == 'bnode': return rdflib.BNode(d['value']) else: raise NotImplementedError("json term type %r" % t) def parseJsonBindings(bindings): newbindings = [] for row in bindings: outRow = {} for k, v in row.items(): outRow[k] = parseJsonTerm(v) newbindings.append(outRow) return newbindings # Helper functions to form join of mutiple binding sets def joinBinding(result_binding, constraint_binding): for k in result_binding: if k in constraint_binding: if result_binding[k] != constraint_binding[k]: return None joined_binding = result_binding.copy() joined_binding.update(constraint_binding) return joined_binding def joinBindings(result_bindings, constraint_bindings): return [ bj for bj in [ joinBinding(b1, b2) for b1 in result_bindings for b2 in constraint_bindings ] if bj ] def joinBindingsToJSON(result_bindings, constraint_bindings): return [ bindingToJSON(bj) for bj in [ joinBinding(b1, b2) for b1 in result_bindings for b2 in constraint_bindings ] if bj ] # Helper functions for accessing data at URI reference, which may be a path relative to current directory def resolveUri(uriref, base, path=""): """ Resolve a URI reference against a supplied base URI and path. (The path is a local file system path, and may need converting to use URI conventions) """ upath = urllib.request.pathname2url(path) if os.path.isdir(path) and not upath.endswith('/'): upath = upath + '/' return urllib.parse.urljoin(urllib.parse.urljoin(base, upath), uriref) def retrieveUri(uriref): uri = resolveUri(uriref, "file://", os.getcwd()) log.debug("retrievUri: %s"%(uri)) request = urllib.request.Request(uri) try: response = io.TextIOWrapper(urllib.request.urlopen(request), encoding="utf-8") result = response.read() except: result = None return result # Helper function for determining type of query def queryType(query): """ Returns "ASK", "SELECT", "CONSTRUCT", "DESCRIBE" or None """ iriregex = "<[^>]*>" baseregex = ".*base.*"+iriregex prefixregex = ".*prefix.*"+iriregex queryregex = "^("+baseregex+")?("+prefixregex+")*.*(ask|select|construct|describe).*$" match = re.match(queryregex, query, flags=re.IGNORECASE|re.DOTALL) if match: return match.group(3).upper() return None # Main program functions def getQuery(options, args): """ Get query string from command line option or argument. """ if options.query: return retrieveUri(options.query) elif len(args) >= 2: return args[1] return None def getPrefixes(options): """ Get prefix string from command line option. """ defaultPrefixes = """ PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#> PREFIX owl: <http://www.w3.org/2002/07/owl#> PREFIX xsd: <http://www.w3.org/2001/XMLSchema#> PREFIX dcterms: <http://purl.org/dc/terms/> PREFIX foaf: <http://xmlns.com/foaf/0.1/> """ # PREFIX xml: <http://www.w3.org/XML/1998/namespace> configbase = os.path.expanduser("~") prefixUri = options.prefix or resolveUri( ".asqc-prefixes", "file://", configbase) if prefixUri.startswith("~"): prefixUri = configbase+prefixUri[1:] log.debug("Prefix URI %s"%(prefixUri)) prefixes = retrieveUri(prefixUri) return prefixes or defaultPrefixes def getBindings(options): bndtext = None bindings = ( { "head": { "vars": [] } , "results": { "bindings": [{}] } }) if options.bindings and options.bindings != "-": bndtext = retrieveUri(options.bindings) elif options.bindings == "-": if options.rdf_data or options.endpoint: bndtext = sys.stdin.read() else: # Can't read bindings from stdin if trying to read RDF from stdin return None else: bndtext = None if bndtext: try: bindings = json.loads(bndtext) bindings['results']['bindings'] = parseJsonBindings(bindings['results']['bindings']) except Exception as e: bindings = None return bindings def getRdfData(options): """ Reads RDF data from files specified using -r or from stdin """ if not options.rdf_data: options.rdf_data = ['-'] rdfgraph = rdflib.Graph() for r in options.rdf_data: base = "" if r == "-": rdftext = sys.stdin.read() else: log.debug("Reading RDF from %s"%(r)) rdftext = retrieveUri(r) base = r rdfformatdefault = RDFTYPPARSERMAP[RDFTYP[0]] rdfformatselect = RDFTYPPARSERMAP.get(options.format_rdf_in, rdfformatdefault) try: log.debug("Parsing RDF format %s"%(rdfformatselect)) if rdfformatselect == "rdfa+html": rdfgraph.parse(data=rdftext, format="rdfa", media_type="text/html", publicID=base) else: rdfgraph.parse(data=rdftext, format=rdfformatselect, publicID=base) except Exception as e: log.debug("RDF Parse failed: %s"%(repr(e))) log.debug("traceback: %s"%(traceback.format_exc())) return None return rdfgraph def queryRdfData(progname, options, prefixes, query, bindings): """ Submit query against RDF data. Result is tuple of status and dictionary/list structure suitable for JSON encoding, or an rdflib.graph value. """ rdfgraph = getRdfData(options) if not rdfgraph: print( "%s: Could not read RDF data, or syntax error in input"%progname ) print( " Use -r <file> or supply RDF on stdin; specify input format if not RDF/XML" ) return (2, None) query = prefixes + query log.debug("queryRdfData query:\n%s\n"%(query)) try: resps = [rdfgraph.query(query, initBindings=b) for b in bindings['results']['bindings']] except AssertionError as e: print( "Query failed (query syntax problem?)" ) print( "Submitted query:" ) print( query ) return (2, None) res = { "head": {} } if resps[0].type == 'ASK': res["boolean"] = any([ r.askAnswer for r in resps ]) return (0 if res["boolean"] else 1, res) elif resps[0].type == 'SELECT': res["head"]["vars"] = resps[0].vars res["results"] = {} res["results"]["bindings"] = [ bindingToJSON(b) for r in resps for b in r.bindings ] return (0 if len(res["results"]["bindings"]) > 0 else 1, res) elif resps[0].type == 'CONSTRUCT': res = rdflib.graph.ReadOnlyGraphAggregate( [r.graph for r in resps] ) return (0 if len(res) > 0 else 1, res) else: assert False, "Unexpected query response type %s"%resp.type return (2, None) def querySparqlEndpoint(progname, options, prefixes, query, bindings): """ Issue SPARQL query to SPARQL HTTP endpoint. Requests either JSON or RDF/XML depending on query type. Returns JSON-like dictionary/list structure or RDF graph, depending on query type. These are used as basis for result formatting by outputResult function """ query = prefixes + query resulttype = "application/RDF+XML" resultjson = False querytype = queryType(query) if querytype in ["ASK", "SELECT"]: # NOTE application/json doesn't work with Fuseki # See: http://gearon.blogspot.co.uk/2011/09/sparql-json-after-commenting-other-day.html resulttype = "application/sparql-results+json" resultjson = True if options.verbose: print( "== Query to endpoint ==" ) print( query ) print( "== resulttype: "+resulttype ) print( "== resultjson: "+str(resultjson) ) sc = SparqlHttpClient(endpointuri=options.endpoint) ((status, reason), result) = sc.doQueryPOST(query, accept=resulttype, JSON=False) if status != 200: assert False, "Error from SPARQL query request: %i %s"%(status, reason) if options.verbose: print( "== Query response ==" ) print( result ) if resultjson: result = json.loads(result) status = 1 if querytype == "SELECT": result['results']['bindings'] = parseJsonBindings(result['results']['bindings']) result['results']['bindings'] = joinBindingsToJSON( result['results']['bindings'], bindings['results']['bindings']) if result['results']['bindings']: status = 0 elif bindings: assert False, "Can't use supplied bindings with endpoint query other than SELECT" elif querytype == "ASK": # Just return JSON from Sparql query if result['boolean']: status = 0 else: # return RDF rdfgraph = rdflib.Graph() try: # Note: declaring xml prefix in SPAQL query can result in invalid XML from Fuseki (v2.1) # See: https://issues.apache.org/jira/browse/JENA-24 rdfgraph.parse(data=result) result = rdfgraph # Return parsed RDF graph if len(result) > 0: status = 0 except Exception as e: assert False, "Error parsing RDF from SPARQL endpoint query: "+str(e) return (status, result) def outputResult(progname, options, result): outstr = sys.stdout if options.output and options.output != "-": print( "Output to other than stdout not implemented" ) if isinstance(result, rdflib.Graph): rdfformatdefault = RDFTYPSERIALIZERMAP[RDFTYP[0]] rdfformatselect = RDFTYPSERIALIZERMAP.get(options.format_rdf_out, rdfformatdefault) result.serialize(destination=outstr, format=rdfformatselect, base=None) elif isinstance(result, str): outstr.write(result) else: if options.format_var_out == "JSON" or options.format_var_out == None: outstr.write(json.dumps(result)) outstr.write("\n") elif options.format_var_out == "XML": writeResultsXML(outstr, result) elif options.format_var_out == "CSV": qvars = result["head"]["vars"] outstr.write(", ".join(qvars)) outstr.write("\n") for bindings in result["results"]["bindings"]: ### print("---- bindings: "+repr(bindings)) vals = [ termToCSV(bindings.get(str(v),{'type': 'literal', 'value': ''})) for v in qvars ] outstr.write(", ".join(vals)) outstr.write("\n") else: for bindings in result["results"]["bindings"]: #log.debug("options.format_var_out '%s'"%(repr(options.format_var_out))) formattedrow = formatBindings(options.format_var_out, bindings) #log.debug("formattedrow '%s'"%(repr(formattedrow))) outstr.write(formattedrow) return def run(configbase, options, args): status = 0 if options.examples: print( "%s/examples"%(os.path.dirname(os.path.abspath(__file__))) ) return 0 progname = os.path.basename(args[0]) query = getQuery(options, args) if not query: print( "%s: Could not determine query string (need query argument or -q option)"%progname ) print( "Run '%s --help' for more information"%progname ) return 2 prefixes = getPrefixes(options) if not prefixes: print( "%s: Could not determine query prefixes"%progname ) print( "Run '%s --help' for more information"%progname ) return 2 ## log.debug("Prefixes:\n%s\n"%(prefixes)) bindings = getBindings(options) if not bindings: print( "%s: Could not determine incoming variable bindings"%progname ) print( "Run '%s --help' for more information"%progname ) return 2 if options.verbose: print( "== Options ==" ) print( repr(options) ) print( "== Prefixes ==" ) print( prefixes ) print( "== Query ==" ) print( query ) print( "== Initial bindings ==" ) print( bindings ) if options.endpoint: (status,result) = querySparqlEndpoint(progname, options, prefixes, query, bindings) else: (status,result) = queryRdfData(progname, options, prefixes, query, bindings) if result: outputResult(progname, options, result) return status def parseCommandArgs(argv): """ Parse command line arguments argv -- argument list from command line Returns a pair consisting of options specified as returned by OptionParser, and any remaining unparsed arguments. """ # create a parser for the command line options parser = optparse.OptionParser( usage=("\n"+ " %prog [options] [query]\n"+ " %prog --help for an options summary\n"+ " %prog --examples to display the path containing example queries"), description="A sparql query client, designed to be used as a filter in a command pipeline. "+ "Pipelined data can be RDF or query variable binding sets, depending on the options used.", version="%prog "+asqc_settings.VERSION) parser.add_option("--examples", action="store_true", dest="examples", default=False, help="display path of examples directory and exit") parser.add_option("-b", "--bindings", dest="bindings", default=None, help="URI or filename of resource containing incoming query variable bindings "+ "(default none). "+ "Specify '-' to use stdin. "+ "This option works for SELECT queries only when accessing a SPARQL endpoint.") parser.add_option("--debug", action="store_true", dest="debug", default=False, help="run with full debug output enabled") parser.add_option("-e", "--endpoint", dest="endpoint", default=None, help="URI of SPARQL endpoint to query.") parser.add_option("-f", "--format", dest="format", default=None, help="Format for input and/or output: "+ "RDFXML, N3, NT, TURTLE, JSONLD, RDFA, HTML5, JSON, CSV or template. "+ "XML, N3, NT, TURTLE, JSONLD, RDFA, HTML5 apply to RDF data, "+ "others apply to query variable bindings. "+ "Multiple comma-separated values may be specified; "+ "they are applied to RDF or variable bindings as appropriate. "+ "'template' is a python formatting template with '%(var)s' for query variable 'var'. "+ "If two values are given for RDF or variable binding data, "+ "they are applied to input and output respectively. "+ "Thus: RDFXML,JSON = RDF/XML and JSON result bindings; "+ "RDFXML,N3 = RDF/XML input and Turtle output; etc.") parser.add_option("-o", "--output", dest="output", default='-', help="URI or filename of RDF resource for output "+ "(default stdout)."+ "Specify '-'to use stdout.") parser.add_option("-p", "--prefix", dest="prefix", default="~/.asqc-prefixes", help="URI or filename of resource containing query prefixes "+ "(default %default)") parser.add_option("-q", "--query", dest="query", help="URI or filename of resource containing query to execute. "+ "If not present, query must be supplied as command line argument.") parser.add_option("-r", "--rdf-input", action="append", dest="rdf_data", default=None, help="URI or filename of RDF resource to query "+ "(default stdin or none). "+ "May be repeated to merge multiple input resources. "+ "Specify '-' to use stdin.") parser.add_option("-v", "--verbose", action="store_true", dest="verbose", default=False, help="display verbose output") parser.add_option("--query-type", dest="query_type", default=None, help="Type of query output: SELECT (variable bindings, CONSTRUCT (RDF) or ASK (status). "+ "May be used when system cannot tell the kind of result by analyzing the query itself. "+ "(Currently not used)") parser.add_option("--format-rdf-in", dest="format_rdf_in", default=None, help="Format for RDF input data: RDFXML, N3, NT, TURTLE, JSONLD, RDFA or HTML5. "+ "RDFA indicates RDFa embedded in XML (or XHTML); "+ "HTML5 indicates RDFa embedded in HTML5.") parser.add_option("--format-rdf-out", dest="format_rdf_out", default=None, help="Format for RDF output data: RDFXML, N3, NT, TURTLE or JSONLD.") parser.add_option("--format-var-in", dest="format_var_in", default=None, help="Format for query variable binding input data: JSON or CSV.") parser.add_option("--format-var-out", dest="format_var_out", default=None, help="Format for query variable binding output data: JSON, CSV or template. "+ "The template option is a Python format string applied to a dictionary of query result variables.") # parse command line now (options, args) = parser.parse_args(argv) if len(args) < 1: parser.error("No command present") if len(args) > 2: parser.error("Too many arguments present: "+repr(args)) def pick_next_format_option(s,kws): t = s for k in kws: if s.upper().startswith(k): s = s[len(k):] if s == "": return (k, "") if s.startswith(','): return (k, s[1:]) break return (t, "") if options.format: fs = options.format while fs: fn,fs = pick_next_format_option(fs, RDFTYP+VARTYP) if fn in RDFTYP: if not options.format_rdf_in: options.format_rdf_in = fn if fn in RDFTYPSERIALIZERMAP: options.format_rdf_out = fn else: if not options.format_var_in and fn in VARTYP: options.format_var_in = fn options.format_var_out = fn if options.verbose: print( "RDF graph input format: "+repr(options.format_rdf_in) ) print( "RDF graph output format: "+repr(options.format_rdf_out) ) print( "Var binding input format: "+repr(options.format_var_in) ) print( "Var binding output format: "+repr(options.format_var_out) ) return (options, args) def runCommand(configbase, argv): """ Run program with supplied configuration base directory, Base directory from which to start looking for research objects, and arguments. This is called by main function (below), and also by test suite routines. Returns exit status. """ log.debug("runCommand: configbase %s, argv %s"%(configbase, repr(argv))) (options, args) = parseCommandArgs(argv) if not options or options.debug: logging.basicConfig(level=logging.DEBUG) status = 2 if options: status = run(configbase, options, args) return status def runMain(): """ Main program transfer function for setup.py console script """ configbase = os.path.expanduser("~") return runCommand(configbase, sys.argv) if __name__ == "__main__": """ Program invoked from the command line. """ # main program status = runMain() sys.exit(status) #--------+---------+---------+---------+---------+---------+---------+---------+
[ "gk-github@ninebynine.org" ]
gk-github@ninebynine.org
610057694e00c3e4fac05320e103e3137f135d00
76192480d7469e3d7f6ac8d8bbc3334445e5fddc
/app.py
07f63fbd87e728aa0ad6e9cd795f03b20816c8e7
[]
no_license
forgeworks/splendor
b7d383a154bf72701a00d005f9aafbd3e90a6b30
f99d66b76971f318637944a8ce5921367ee4aa21
refs/heads/master
2023-05-12T03:07:17.860147
2020-04-03T17:38:55
2020-04-03T17:38:55
155,748,967
0
0
null
null
null
null
UTF-8
Python
false
false
119
py
from flask import Flask from api_example import GreetingV1 app = Flask(__name__) app.register_blueprint(GreetingV1())
[ "deadwisdom@gmail.com" ]
deadwisdom@gmail.com
c93ba3313bf6c3ee32e36cad9c787f55c5d4548b
8395ffb48750359d1bd51a201a41c7fe124998bc
/apc2015/perception/single_utils/src/generate_naive_cloud.py
4195bb9783faaf79d4485ed09ada91429266c3d6
[]
no_license
duke-iml/ece490-s2016
ab6c3d3fb159a28a9c38487cdb1ad3993008b854
f9cc992fbaadedc8a69678ba39f0c9d108e6910d
refs/heads/master
2020-04-12T09:03:56.601000
2016-11-29T21:36:48
2016-11-29T21:36:48
49,226,568
2
6
null
2016-11-29T21:36:49
2016-01-07T19:42:34
Python
UTF-8
Python
false
false
2,659
py
#!/usr/bin/env python from __future__ import division import sys import rospy import cv2 from std_msgs.msg import String from sensor_msgs.msg import Image from cv_bridge import CvBridge, CvBridgeError import subprocess import time import psutil import sys import os import matplotlib.pyplot as plt import numpy as np from common_utils import * from math import pi, sin, cos, tan, atan, sqrt pid = None file_name = None rgb_mat = None depth_mat = None bridge = CvBridge() w = 320 h = 240 diag_ang = 74/180*pi diag = sqrt(w**2+h**2) lift = diag/2 / tan(diag_ang/2) def receive_rgb(data): global rgb_mat rgb_mat = bridge.imgmsg_to_cv2(data, "bgr8") if depth_mat is not None: process() def receive_depth(data): global depth_mat depth_mat = bridge.imgmsg_to_cv2(data, "mono16") depth_mat = depth_mat[:,:,0] if rgb_mat is not None: process() def process(): psutil.Process(pid).kill() cv2.imwrite(file_name+".bmp", rgb_mat) cv2.imwrite(file_name+".depth.bmp", depth_mat) assert depth_mat.shape == (h, w) point_cloud = [] for i in range(h): for j in range(w): depth = depth_mat[i, j] b1, g1, r1 = list(rgb_mat[i*2, j*2, :].flatten()) b2, g2, r2 = list(rgb_mat[i*2+1, j*2, :].flatten()) b3, g3, r3 = list(rgb_mat[i*2, j*2+1, :].flatten()) b4, g4, r4 = list(rgb_mat[i*2+1, j*2+1, :].flatten()) b1 = int(b1) b2 = int(b2) b3 = int(b3) b4 = int(b4) g1 = int(g1) g2 = int(g2) g3 = int(g3) g4 = int(g4) r1 = int(r1) r2 = int(r2) r3 = int(r3) r4 = int(r4) r = int((r1+r2+r3+r4)/4) g = int((g1+g2+g3+g4)/4) b = int((b1+b2+b3+b4)/4) rgb = rgb_to_pcl_float(r1, g1, b1) if depth==32001: continue assert depth<20000 coord = (j+0.5-w/2, i+0.5-h/2) real_x = coord[0]/lift*depth real_y = coord[1]/lift*depth point_cloud.append([real_x/1000, real_y/1000, depth/1000, rgb]) write_pcd_file(point_cloud, file_name) rospy.signal_shutdown("Point cloud made, shutting down...\n") def main(): global file_name if len(sys.argv)>=2: file_name = sys.argv[1] else: file_name = 'point_cloud.pcd' global pid process = subprocess.Popen('hardware_layer/RealSense_ROS_Emitter', stdout=subprocess.PIPE, stderr=subprocess.PIPE) pid = process.pid time.sleep(3) rospy.init_node('naive_point_cloud', disable_signals=True) rgb_sub = rospy.Subscriber("/realsense/rgb", Image, receive_rgb, queue_size=1) depth_sub = rospy.Subscriber("/realsense/depth", Image, receive_depth, queue_size=1) rospy.spin() if __name__ == '__main__': main()
[ "hauser.kris@gmail.com" ]
hauser.kris@gmail.com
68062176f60959a3d7bd5ed6cfbb91d826843649
ca5b5c217e0053645c2664d777699e9a5050715e
/python/primitive.py
1352c69b984be863d2167845164d2ffaa39c93e0
[ "MIT" ]
permissive
rodluger/starrynight
1405ffdb5a0dd0fefc0ae34e7cdaf7eab4735356
d3f015e466621189cb271d4d18b538430b14a557
refs/heads/master
2021-10-26T03:32:15.220725
2021-10-22T15:16:48
2021-10-22T15:16:48
236,542,672
7
1
MIT
2020-06-03T19:51:10
2020-01-27T16:58:05
Jupyter Notebook
UTF-8
Python
false
false
15,076
py
from special import hyp2f1, J, ellip from utils import * from vieta import Vieta from linear import dP2 import matplotlib.pyplot as plt import numpy as np __ALL__ = ["compute_P", "compute_Q", "comput_T"] def compute_U(vmax, s1): """ Given s1 = sin(0.5 * kappa), compute the integral of cos(x) sin^v(x) from 0.5 * kappa1 to 0.5 * kappa2 recursively and return an array containing the values of this function from v = 0 to v = vmax. """ U = np.empty(vmax + 1) U[0] = pairdiff(s1) term = s1 ** 2 for v in range(1, vmax + 1): U[v] = pairdiff(term) / (v + 1) term *= s1 return U def compute_I(nmax, kappa, s1, c1): # Lower boundary I = np.empty(nmax + 1) I[0] = 0.5 * pairdiff(kappa) # Recurse upward s2 = s1 ** 2 term = s1 * c1 for v in range(1, nmax + 1): I[v] = (1.0 / (2 * v)) * ((2 * v - 1) * I[v - 1] - pairdiff(term)) term *= s2 return I def _compute_W_indef(nmax, s2, q2, q3): """ Compute the expression s^(2n + 2) (3 / (n + 1) * 2F1(-1/2, n + 1, n + 2, 1 - q^2) + 2q^3) / (2n + 5) evaluated at n = [0 .. nmax], where s = sin(1/2 kappa) q = (1 - s^2 / k^2)^1/2 by either upward recursion (stable for |1 - q^2| > 1/2) or downward recursion (always stable). """ W = np.empty(nmax + 1) if np.abs(1 - q2) < 0.5: # Setup invs2 = 1 / s2 z = (1 - q2) * invs2 s2nmax = s2 ** nmax x = q2 * q3 * s2nmax # Upper boundary condition W[nmax] = ( s2 * s2nmax * (3 / (nmax + 1) * hyp2f1(-0.5, nmax + 1, nmax + 2, 1 - q2) + 2 * q3) / (2 * nmax + 5) ) # Recurse down for b in range(nmax - 1, -1, -1): f = 1 / (b + 1) A = z * (1 + 2.5 * f) B = x * f W[b] = A * W[b + 1] + B x *= invs2 else: # Setup z = s2 / (1 - q2) x = -2 * q3 * (z - s2) * s2 # Lower boundary condition W[0] = (2 / 5) * (z * (1 - q3) + s2 * q3) # Recurse up for b in range(1, nmax + 1): f = 1 / (2 * b + 5) A = z * (2 * b) * f B = x * f W[b] = A * W[b - 1] + B x *= s2 return W def compute_W(nmax, s2, q2, q3): return pairdiff( np.array([_compute_W_indef(nmax, s2[i], q2[i], q3[i]) for i in range(len(s2))]) ) def compute_J(nmax, k2, km2, kappa, s1, s2, c1, q2, dF, dE): """ Return the array J[0 .. nmax], computed recursively using a tridiagonal solver and a lower boundary condition (analytic in terms of elliptic integrals) and an upper boundary condition (computed numerically). """ # Boundary conditions z = s1 * c1 * np.sqrt(q2) resid = km2 * pairdiff(z) f0 = (1 / 3) * (2 * (2 - km2) * dE + (km2 - 1) * dF + resid) fN = J(nmax, k2, kappa) # Set up the tridiagonal problem a = np.empty(nmax - 1) b = np.empty(nmax - 1) c = np.empty(nmax - 1) term = k2 * z * q2 ** 2 for i, v in enumerate(range(2, nmax + 1)): amp = 1.0 / (2 * v + 3) a[i] = -2 * (v + (v - 1) * k2 + 1) * amp b[i] = (2 * v - 3) * k2 * amp c[i] = pairdiff(term) * amp term *= s2 # Add the boundary conditions c[0] -= b[0] * f0 c[-1] -= fN # Construct the tridiagonal matrix A = np.diag(a, 0) + np.diag(b[1:], -1) + np.diag(np.ones(nmax - 2), 1) # Solve soln = np.linalg.solve(A, c) return np.concatenate(([f0], soln, [fN])) def K(I, delta, u, v): """Return the integral K, evaluated as a sum over I.""" return sum([Vieta(i, u, v, delta) * I[i + u] for i in range(u + v + 1)]) def L(J, k, delta, u, v, t): """Return the integral L, evaluated as a sum over J.""" return k ** 3 * sum( [Vieta(i, u, v, delta) * J[i + u + t] for i in range(u + v + 1)] ) def compute_H(uvmax, xi, gradient=False): c = np.cos(xi) s = np.sin(xi) cs = c * s cc = c ** 2 ss = s ** 2 H = np.empty((uvmax + 1, uvmax + 1)) dH = np.empty((uvmax + 1, uvmax + 1, len(xi))) H[0, 0] = pairdiff(xi) dH[0, 0] = 1 H[1, 0] = pairdiff(s) dH[1, 0] = c H[0, 1] = -pairdiff(c) dH[0, 1] = s H[1, 1] = -0.5 * pairdiff(cc) dH[1, 1] = cs for u in range(2): for v in range(2, uvmax + 1 - u): H[u, v] = (-pairdiff(dH[u, v - 2] * cs) + (v - 1) * H[u, v - 2]) / (u + v) dH[u, v] = dH[u, v - 2] * ss for u in range(2, uvmax + 1): for v in range(uvmax + 1 - u): H[u, v] = (pairdiff(dH[u - 2, v] * cs) + (u - 1) * H[u - 2, v]) / (u + v) dH[u, v] = dH[u - 2, v] * cc if gradient: return H, dH else: return H def _compute_T2_indef(b, xi): """ Note: requires b >= 0. """ s = np.sin(xi) c = np.cos(xi) t = s / c sgn = np.sign(s) bc = np.sqrt(1 - b ** 2) bbc = b * bc # Special cases if xi == 0: return -(np.arctan((2 * b ** 2 - 1) / (2 * bbc)) + bbc) / 3 elif xi == 0.5 * np.pi: return (0.5 * np.pi - np.arctan(b / bc)) / 3 elif xi == np.pi: return (0.5 * np.pi + bbc) / 3 elif xi == 1.5 * np.pi: return (0.5 * np.pi + np.arctan(b / bc) + 2 * bbc) / 3 # Figure out the offset if xi < 0.5 * np.pi: delta = 0 elif xi < np.pi: delta = np.pi elif xi < 1.5 * np.pi: delta = 2 * bbc else: delta = np.pi + 2 * bbc # We're done return ( np.arctan(b * t) - sgn * (np.arctan(((s / (1 + c)) ** 2 + 2 * b ** 2 - 1) / (2 * bbc)) + bbc * c) + delta ) / 3 def compute_P(ydeg, bo, ro, kappa): """Compute the P integral.""" # Basic variables delta = (bo - ro) / (2 * ro) k2 = (1 - ro ** 2 - bo ** 2 + 2 * bo * ro) / (4 * bo * ro) k = np.sqrt(k2) km2 = 1.0 / k2 fourbr15 = (4 * bo * ro) ** 1.5 k3fourbr15 = k ** 3 * fourbr15 tworo = np.empty(ydeg + 4) tworo[0] = 1.0 for i in range(1, ydeg + 4): tworo[i] = tworo[i - 1] * 2 * ro # Pre-compute the helper integrals x = 0.5 * kappa s1 = np.sin(x) s2 = s1 ** 2 c1 = np.cos(x) q2 = 1 - np.minimum(1.0, s2 / k2) q3 = q2 ** 1.5 U = compute_U(2 * ydeg + 5, s1) I = compute_I(ydeg + 3, kappa, s1, c1) W = compute_W(ydeg, s2, q2, q3) # Compute the elliptic integrals F, E, PIprime = ellip(bo, ro, kappa) J = compute_J(ydeg + 1, k2, km2, kappa, s1, s2, c1, q2, F, E) # Now populate the P array P = np.zeros((ydeg + 1) ** 2) n = 0 for l in range(ydeg + 1): for m in range(-l, l + 1): mu = l - m nu = l + m if (mu / 2) % 2 == 0: # Same as in starry P[n] = 2 * tworo[l + 2] * K(I, delta, (mu + 4) // 4, nu // 2) elif mu == 1: if l == 1: # Same as in starry, but using expression from Pal (2012) P[2] = dP2(bo, ro, k2, kappa, s1, s2, c1, F, E, PIprime) elif l % 2 == 0: # Same as in starry P[n] = ( tworo[l - 1] * fourbr15 * ( L(J, k, delta, (l - 2) // 2, 0, 0) - 2 * L(J, k, delta, (l - 2) // 2, 0, 1) ) ) else: # Same as in starry P[n] = ( tworo[l - 1] * fourbr15 * ( L(J, k, delta, (l - 3) // 2, 1, 0) - 2 * L(J, k, delta, (l - 3) // 2, 1, 1) ) ) elif (mu - 1) / 2 % 2 == 0: # Same as in starry P[n] = ( 2 * tworo[l - 1] * fourbr15 * L(J, k, delta, (mu - 1) // 4, (nu - 1) // 2, 0) ) else: """ A note about these cases. In the original starry code, these integrals are always zero because the integrand is antisymmetric about the midpoint. Now, however, the integration limits are different, so there's no cancellation in general. The cases below are just the first and fourth cases in equation (D25) of the starry paper. We can re-write them as the first and fourth cases in (D32) and (D35), respectively, but note that we pick up a factor of `sgn(cos(phi))`, since the power of the cosine term in the integrand is odd. The other thing to note is that `u` in the call to `K(u, v)` is now a half-integer, so our Vieta trick (D36, D37) doesn't work out of the box. """ if nu % 2 == 0: res = 0 u = int((mu + 4.0) // 4) v = int(nu / 2) for i in range(u + v + 1): res += Vieta(i, u, v, delta) * U[2 * (u + i) + 1] P[n] = 2 * tworo[l + 2] * res else: res = 0 u = (mu - 1) // 4 v = (nu - 1) // 2 for i in range(u + v + 1): res += Vieta(i, u, v, delta) * W[i + u] P[n] = tworo[l - 1] * k3fourbr15 * res n += 1 return P def compute_Q(ydeg, lam, gradient=False): # Pre-compute H if gradient: H, dH = compute_H(ydeg + 2, lam, gradient=True) else: H = compute_H(ydeg + 2, lam) # Allocate Q = np.zeros((ydeg + 1) ** 2) dQdlam = np.zeros(((ydeg + 1) ** 2, len(lam))) # Note that the linear term is special Q[2] = pairdiff(lam) / 3 dQdlam[2] = np.ones_like(lam) / 3 # Easy! n = 0 for l in range(ydeg + 1): for m in range(-l, l + 1): mu = l - m nu = l + m if nu % 2 == 0: Q[n] = H[(mu + 4) // 2, nu // 2] if gradient: dQdlam[n] = dH[(mu + 4) // 2, nu // 2] n += 1 # Enforce alternating signs for (lower, upper) limits dQdlam *= np.repeat([-1, 1], len(lam) // 2).reshape(1, -1) if gradient: return Q, dQdlam else: return Q def compute_T(ydeg, b, theta, xi): # Pre-compute H H = compute_H(ydeg + 2, xi) # Vars ct = np.cos(theta) st = np.sin(theta) ttinvb = st / (b * ct) invbtt = ct / (b * st) b32 = (1 - b ** 2) ** 1.5 bct = b * ct bst = b * st # Recurse T = np.zeros((ydeg + 1) ** 2) # Case 2 (special) T[2] = pairdiff([np.sign(b) * _compute_T2_indef(np.abs(b), x) for x in xi]) # Special limit: sin(theta) = 0 if np.abs(st) < STARRY_T_TOL: sgnct = np.sign(ct) n = 0 for l in range(ydeg + 1): for m in range(-l, l + 1): mu = l - m nu = l + m if nu % 2 == 0: T[n] = sgnct ** l * b ** (1 + nu // 2) * H[(mu + 4) // 2, nu // 2] else: if mu == 1: if (l % 2) == 0: T[n] = -sgnct * b32 * H[l - 2, 4] elif l > 1: T[n] = -b * b32 * H[l - 3, 5] else: T[n] = sgnct ** (l - 1) * ( b32 * b ** ((nu + 1) // 2) * H[(mu - 1) // 2, (nu + 5) // 2] ) n += 1 return T # Special limit: cos(theta) = 0 elif np.abs(ct) < STARRY_T_TOL: sgnst = np.sign(st) n = 0 for l in range(ydeg + 1): for m in range(-l, l + 1): mu = l - m nu = l + m if nu % 2 == 0: T[n] = b ** ((mu + 2) // 2) * H[nu // 2, (mu + 4) // 2] if sgnst == 1: T[n] *= (-1) ** (mu // 2) else: T[n] *= (-1) ** (nu // 2) else: if mu == 1: if (l % 2) == 0: T[n] = ( (-sgnst) ** (l - 1) * b ** (l - 1) * b32 * H[1, l + 1] ) elif l > 1: T[n] = b ** (l - 2) * b32 * H[2, l] if sgnst == 1: T[n] *= (-1) ** l else: T[n] *= -1 else: T[n] = ( b32 * b ** ((mu - 3) // 2) * H[(nu - 1) // 2, (mu + 5) // 2] ) if sgnst == 1: T[n] *= (-1) ** ((mu - 1) // 2) else: T[n] *= (-1) ** ((nu - 1) // 2) n += 1 return T # Cases 1 and 5 jmax = 0 Z0 = 1 for nu in range(0, 2 * ydeg + 1, 2): kmax = 0 Z1 = Z0 for mu in range(0, 2 * ydeg - nu + 1, 2): l = (mu + nu) // 2 n1 = l ** 2 + nu n5 = (l + 2) ** 2 + nu + 1 Z2 = Z1 for j in range(jmax + 1): Z_1 = -bst * Z2 Z_5 = b32 * Z2 for k in range(kmax + 1): p = j + k q = l + 1 - (j + k) fac = -invbtt / (k + 1) T[n1] += Z_1 * (bct * H[p + 1, q] - st * H[p, q + 1]) Z_1 *= (kmax + 1 - k) * fac if n5 < (ydeg + 1) ** 2: T[n5] += Z_5 * (bct * H[p + 1, q + 2] - st * H[p, q + 3]) Z_5 *= (kmax - k) * fac T[n1] += Z_1 * (bct * H[p + 2, q - 1] - st * H[p + 1, q]) Z2 *= (jmax - j) / (j + 1) * ttinvb kmax += 1 Z1 *= -bst jmax += 1 Z0 *= bct # Cases 3 and 4 Z0 = b32 kmax = 0 for l in range(2, ydeg + 1, 2): n3 = l ** 2 + 2 * l - 1 n4 = (l + 1) ** 2 + 2 * l + 1 Z = Z0 for k in range(kmax + 1): p = k q = l + 1 - k T[n3] -= Z * (bst * H[p + 1, q] + ct * H[p, q + 1]) if l < ydeg: T[n4] -= Z * ( bst * st * H[p + 2, q] + bct * ct * H[p, q + 2] + (1 + b ** 2) * st * ct * H[p + 1, q + 1] ) Z *= -(kmax - k) / (k + 1) * invbtt kmax += 2 Z0 *= bst ** 2 return T
[ "rodluger@gmail.com" ]
rodluger@gmail.com
1a6991bb0eca5e3ab71d0ba1a68d1884bc4012b5
8778be5676e3657be3db6d71bb81b57ef8020dae
/lab2/webcam_prediction.py
e723c7d678f951f7d107b5fe986ae7fd7900e902
[]
no_license
JerelynCo/Pattern-Recognition
ab922e35c36fb51d2d1c2996d2483d5b152c471e
ff36571abf8c12b5f9bae7e5dd89afcce35f5e1e
refs/heads/master
2016-08-12T20:16:24.736727
2016-04-19T05:40:52
2016-04-19T05:40:52
53,573,471
1
0
null
null
null
null
UTF-8
Python
false
false
2,886
py
import numpy as np import cv2 import math import json def calculateProbability(x, mean, stdev): exponent = math.exp(-(math.pow(x - mean, 2) / (2 * math.pow(stdev, 2)))) return (1 / (math.sqrt(2 * math.pi) * stdev)) * exponent def calculateClassProbabilities(summaries, inputVector): probabilities = {} for classValue, classSummaries in summaries.items(): probabilities[classValue] = 1 for i in range(len(classSummaries)): mean, stdev = classSummaries[i] x = inputVector[i] probabilities[classValue] *= calculateProbability(x, mean, stdev) return probabilities def predict(summaries, inputVector): probabilities = calculateClassProbabilities(summaries, inputVector) bestLabel, bestProb = None, -1 for classValue, probability in probabilities.items(): if bestLabel is None or probability > bestProb: bestProb = probability bestLabel = classValue return bestLabel def auto_canny(image, sigma=0.33): # compute the median of the single channel pixel intensities v = np.median(image) # apply automatic Canny edge detection using the computed median lower = int(max(0, (1.0 - sigma) * v)) upper = int(min(255, (1.0 + sigma) * v)) edged = cv2.Canny(image, lower, upper) # return the edged image return edged cap = cv2.VideoCapture(0) while(True): # Capture frame-by-frame ret, frame = cap.read() # Our operations on the frame come here gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY) blurred = cv2.GaussianBlur(gray, (5, 5), 0) edges = auto_canny(blurred) contoured_image, contours, hierarchy = cv2.findContours( edges, cv2.RETR_EXTERNAL, cv2.CHAIN_APPROX_SIMPLE) # ddepth for sobel ddepth = cv2.CV_8U sobel_x = cv2.Sobel(blurred, ddepth, 1, 0) sobel_y = cv2.Sobel(blurred, ddepth, 1, 0) sobel = cv2.addWeighted(sobel_x, 0.5, sobel_y, 0.5, 0) counter = 0 image_out = frame.copy() for contour in contours: # get box bounding contour [x, y, w, h] = cv2.boundingRect(contour) if h < image_out.shape[1] * 0.05 or w < image_out.shape[0] * 0.05: continue crop_img_sobel = sobel[y:y + h, x:x + w] resized = cv2.resize(crop_img_sobel, (8, 8)).flatten() / 255 with open('classifier/train_summary.json', 'r') as f: train_summary = json.loads(f.read()) if(predict(train_summary, resized) == '1'): cv2.rectangle(edges, (x, y), (x + w, y + h), (255, 0, 0), 2) else: cv2.rectangle(edges, (x, y), (x + w, y + h), (0, 0, 255), 2) counter += 1 # Display the resulting frame cv2.imshow('frame', edges) if cv2.waitKey(1) & 0xFF == ord('q'): break # When everything done, release the capture cap.release() cv2.destroyAllWindows()
[ "hadrianpaulo@gmail.com" ]
hadrianpaulo@gmail.com
21c03646d3b3e6dae55e626e2eb651912903a68a
e0d46d7d1048648836af5f5243adc5e84e391382
/forwardgram.py
b540d98db563934ac11526cad28fe18b551ecefb
[]
no_license
DanilenkoDanil/forward
a2ca0a96aa826754f1c23caa016fc70d3eaf5fe1
8a0b5d2f0cadf22990a13fdddb23ce5d0a95a338
refs/heads/main
2023-08-07T21:43:51.561629
2021-09-22T23:18:13
2021-09-22T23:18:13
329,336,856
0
0
null
null
null
null
UTF-8
Python
false
false
1,870
py
from telethon import TelegramClient, events, sync from telethon.tl.types import InputChannel import yaml import sys import logging logging.basicConfig(level=logging.INFO, format='%(asctime)s - %(name)s - %(levelname)s - %(message)s') logging.getLogger('telethon').setLevel(level=logging.WARNING) logger = logging.getLogger(__name__) def start(config): client = TelegramClient(config["session_name"], config["api_id"], config["api_hash"]) client.start() input_channels_entities = [] output_channel_entities = [] for d in client.iter_dialogs(): if d.name in config["input_channel_names"]: input_channels_entities.append(InputChannel(d.entity.id, d.entity.access_hash)) if d.name in config["output_channel_names"]: output_channel_entities.append(InputChannel(d.entity.id, d.entity.access_hash)) if not output_channel_entities: logger.error(f"Could not find any output channels in the user's dialogs") sys.exit(1) if not input_channels_entities: logger.error(f"Could not find any input channels in the user's dialogs") sys.exit(1) logging.info( f"Listening on {len(input_channels_entities)} channels. Forwarding messages to {len(output_channel_entities)} channels.") @client.on(events.NewMessage(chats=input_channels_entities)) async def handler(event): for output_channel in output_channel_entities: await client.forward_messages(output_channel, event.message) client.run_until_disconnected() if __name__ == "__main__": if len(sys.argv) < 2: print(f"Usage: {sys.argv[0]} {{CONFIG_PATH}}") sys.exit(1) with open(sys.argv[1], 'rb') as f: config = yaml.safe_load(f) start(config)
[ "noreply@github.com" ]
noreply@github.com
12d670bebcfaf06a04d1a065f01c5f4d66c7390e
71ae0df6d2c10f2c37c40f527a78716edb23275e
/main.spec
3188163d8287f3b0768c29b22de3a711232c7c87
[]
no_license
MediaNik5/CannonPy
b1637aabe9fe45441fa27515cd7eef649a175f08
0390c4aebb2a3c56073fb490fe5d350cfe995d02
refs/heads/master
2023-03-18T12:30:22.621039
2021-03-10T13:31:06
2021-03-10T13:31:06
346,366,112
0
0
null
null
null
null
UTF-8
Python
false
false
866
spec
# -*- mode: python ; coding: utf-8 -*- block_cipher = None a = Analysis(['main.py'], pathex=['C:\\gdrive\\Works\\projects\\python\\Cannon'], binaries=[], datas=[], hiddenimports=[], hookspath=[], runtime_hooks=[], excludes=[], win_no_prefer_redirects=False, win_private_assemblies=False, cipher=block_cipher, noarchive=False) pyz = PYZ(a.pure, a.zipped_data, cipher=block_cipher) exe = EXE(pyz, a.scripts, a.binaries, a.zipfiles, a.datas, [], name='main', debug=False, bootloader_ignore_signals=False, strip=False, upx=True, upx_exclude=[], runtime_tmpdir=None, console=True )
[ "stafilopok@mail.ru" ]
stafilopok@mail.ru
f0d08ba48ad4eb4e0d09ebaa588e70674bf6d362
3248dd56d1b6ed34eec801a16752328a6016e9fe
/SentimentAnalysis/asgi.py
83b2a88105aacd9830a3a04bad7f6ff45a8787db
[]
no_license
shamil-t/sentiment-analysis-imdb-django
d17905e3fdcde770ba987e37272bd0a4047b3e79
fea9349bd3cc55cf550920e65a53254e23aff20b
refs/heads/main
2023-03-17T01:14:41.376834
2021-02-27T03:36:02
2021-02-27T03:36:02
342,748,491
1
0
null
null
null
null
UTF-8
Python
false
false
411
py
""" ASGI config for SentimentAnalysis project. It exposes the ASGI callable as a module-level variable named ``application``. For more information on this file, see https://docs.djangoproject.com/en/3.1/howto/deployment/asgi/ """ import os from django.core.asgi import get_asgi_application os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'SentimentAnalysis.settings') application = get_asgi_application()
[ "tshamil90@gmail.com" ]
tshamil90@gmail.com
55ba0c58bb1c484542456f12df2c5861a5e9826e
28c155fc7f808365d2f37fff2c794c3994c89353
/learning_web_2_bug/contact/migrations/0001_initial.py
7351287297948ce7a16346e6139964d1f4dcbab8
[]
no_license
bwtzxcvb/rockfracturegroup
97e454ddda1b4407b835c7ef10da6596c53356d3
4e1b61d30928cea82c26ee2d81e8e8b8e391abe5
refs/heads/master
2023-04-27T18:58:01.901009
2019-10-24T10:42:47
2019-10-24T10:42:47
211,757,570
0
0
null
2023-04-21T20:38:23
2019-09-30T02:14:39
Python
UTF-8
Python
false
false
764
py
# Generated by Django 2.2.4 on 2019-09-10 12:47 from django.db import migrations, models class Migration(migrations.Migration): initial = True dependencies = [ ] operations = [ migrations.CreateModel( name='ContactMessage', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('name', models.CharField(max_length=30)), ('email', models.EmailField(max_length=254)), ('message', models.TextField()), ('message_date', models.DateTimeField(auto_now_add=True)), ], options={ 'ordering': ('-message_date',), }, ), ]
[ "1927552890@qq.com" ]
1927552890@qq.com
1ffc8b3649921a8cf943112df31655726ca74210
de24f83a5e3768a2638ebcf13cbe717e75740168
/moodledata/vpl_data/143/usersdata/210/62277/submittedfiles/av2_p3_m2.py
26e0d4120d107718e01aaba735ca255a96ae8f9d
[]
no_license
rafaelperazzo/programacao-web
95643423a35c44613b0f64bed05bd34780fe2436
170dd5440afb9ee68a973f3de13a99aa4c735d79
refs/heads/master
2021-01-12T14:06:25.773146
2017-12-22T16:05:45
2017-12-22T16:05:45
69,566,344
0
0
null
null
null
null
UTF-8
Python
false
false
400
py
# -*- coding: utf-8 -*- def degrais(a): soma=0 degrau=0 for i in range(0,len(a)-1,1): soma=(a[i]-a[i+1]) if soma<0: soma=soma*(-1) if soma>degrau: degrau=soma return degrau h=int(input('digite o valor de h:')) j=[] for i in range(0,h,1): numero=int(input('digite o numero:')) j.append(numero) x=degrais(j) print(x)
[ "rafael.mota@ufca.edu.br" ]
rafael.mota@ufca.edu.br
7c9495dcac44bea7cf3f3cf9a7bd33f81a983d2b
ff3b670a0f618a95d957b813b2607b21947e3dfd
/virtual/lib/python3.7/encodings/koi8_r.py
3a393b48573bf829d26d6512141cdf0ac75fe1b0
[]
no_license
iguzmanl/CSC-325-Project1
2391f28c5b88136daae631bc806db9c5f075ddf7
4c636aad8d7cac5e8f261bdd3b46cf55d77083d7
refs/heads/master
2020-04-20T22:15:29.923159
2019-03-04T05:40:23
2019-03-04T05:40:23
169,133,888
0
0
null
null
null
null
UTF-8
Python
false
false
62
py
/Users/sabaramadan/anaconda3/lib/python3.7/encodings/koi8_r.py
[ "sabaramadan@Sabas-MacBook-Air.local" ]
sabaramadan@Sabas-MacBook-Air.local
ccb00e1329cbbb5be30d2038403c2e5e4303ba75
48660fe39e8b6c49df12bde91e01710f20782f00
/resources/routes/send_qr.py
9cc0c022cd5ff0c29dd33ebbf98d16415396e413
[]
no_license
cringeburger/smart-city-backend
2ed92de8858cf09642f962d323106b52c183f5ab
d4a186d05a91eae802a93eba9981527e1b6432cd
refs/heads/master
2023-04-28T14:26:12.884626
2021-05-23T20:07:38
2021-05-23T20:07:38
369,588,488
0
0
null
null
null
null
UTF-8
Python
false
false
509
py
from flask import request, send_file from resources import app from resources.modules.qr_generator import generate_qr from datetime import date @app.get('/send_qr') def send_qr(): user_token = request.args['user_token'] # mail_domen = request.args['mail_domen'] # subject = request.args['subject'] filename = 'qr_'+str(date.today()) + '_' + user_token generate_qr('https://www.youtube.com/watch?v=dQw4w9WgXcQ', filename) return send_file('generated_qr\\' + filename+ '.png', mimetype='image/png')
[ "istrebitel.3.12@gmail.com" ]
istrebitel.3.12@gmail.com
876ca7ebfa0fdf067bdeaee9c405205dda120f80
fd5010485620821c32e5d2332ff7e4ef0b824dfe
/docs/study/zh-cn/study/QuecPythonTest/code/01_LED.py
80821f9f2b9f0285a3d728d004a4b991497cc57a
[ "MIT" ]
permissive
jks-liu/wiki
cfc1d1abbfc667dd461c40fe25f9ce96bdb2f8fd
397b33f00d128b4b8337a750df59e20356831802
refs/heads/main
2023-02-23T08:49:12.092973
2021-01-27T00:08:39
2021-01-27T00:08:39
333,250,497
0
0
MIT
2021-01-26T23:56:55
2021-01-26T23:56:54
null
UTF-8
Python
false
false
1,967
py
# 实验1: 跑马灯 # API资料参考连接: https://python.quectel.com/wiki/#/zh-cn/api/?id=pin from machine import Pin import utime IOdictRead = {} # 记录已经初始化的GPIO口 IOdictWrite = {} # 记录已经初始化的GPIO口 def GPIO_Read(gpioX, Pull=Pin.PULL_DISABLE, level=1): if IOdictWrite.get(gpioX, None): del IOdictWrite[gpioX] gpioIO = IOdictRead.get(gpioX, None) if gpioIO: return gpioIO.read() else: IOdictRead[gpioX] = (Pin(gpioX, Pin.IN, Pull, level)) gpioIO = IOdictRead.get(gpioX, None) return gpioIO.read() def GPIO_Write(gpioX, level, Pull=Pin.PULL_DISABLE): if IOdictRead.get(gpioX, None): del IOdictRead[gpioX] gpioIO = IOdictWrite.get(gpioX, None) if gpioIO: gpioIO.write(level) else: IOdictWrite[gpioX] = (Pin(gpioX, Pin.OUT, Pull, level)) gpioIO = IOdictWrite.get(gpioX, None) gpioIO.write(level) LED1 = Pin.GPIO1 # 定义LED引脚 LED2 = Pin.GPIO2 # 定义LED引脚 LED3 = Pin.GPIO3 # 定义LED引脚 LED4 = Pin.GPIO4 # 定义LED引脚 LED5 = Pin.GPIO5 # 定义LED引脚 def IO_On(gpioX): # 某个引脚置0 GPIO_Write(gpioX, 0) # 调用写函数 def IO_Off(gpioX): # 某个引脚置1 GPIO_Write(gpioX, 1) # 调用写函数 def IO_All_Off(): # 全部引脚置1 IO_Off(LED1) IO_Off(LED2) IO_Off(LED3) IO_Off(LED4) IO_Off(LED5) def main(): while True: IO_All_Off() # 灭 IO_On(LED1) # 亮 utime.sleep_ms(200) # 延时 IO_All_Off() # 灭 IO_On(LED2) # 亮 utime.sleep_ms(200) # 延时 IO_All_Off() # 灭 IO_On(LED3) # 亮 utime.sleep_ms(200) # 延时 IO_All_Off() # 灭 IO_On(LED4) # 亮 utime.sleep_ms(200) # 延时 IO_All_Off() # 灭 IO_On(LED5) # 亮 utime.sleep_ms(200) # 延时 if __name__ == "__main__": main()
[ "rivern.yuan@quectel.com" ]
rivern.yuan@quectel.com
b12d7388d56385a38eb8484599d6e6bfe729eeb6
00fa9db88ddf3cb70bc33a54e28b7ffe9d9bf42f
/scripts/generate-split-params.py
6a822851c12799fb262cd32c86f22f180b1a0316
[]
no_license
Sixtease/cz-parliament-speech-corpus
445cc1ed2b9e73f1b406a8ab41b95a9b96a5a8f1
e6ba8171ebef8b3f9d8167c3f8561650f43b3130
refs/heads/master
2021-05-21T04:23:37.875930
2020-11-23T18:58:43
2020-11-23T18:58:43
252,540,847
1
0
null
null
null
null
UTF-8
Python
false
false
2,025
py
#!/usr/bin/env python """ first argument is file with one split point in seconds per line on stdin is alignment data in tab-separated format audio start audio end guessed silence length predicted word gold-standard word gold-standard line number prediction - gold-standard match output are tab-separated fields: audio start audio end starting line end line """ import sys import numpy as np splitsfn = sys.argv[1] splitsfh = open(splitsfn, 'r', encoding = 'UTF-8') splits = [float(x.rstrip()) for x in splitsfh.readlines()] splitsfh.close() def init_block(): return { 'matches': [], 'startlineno': None, } def process_block(block): if block['matches'][ 0] < 0.5: sys.stderr.write('discard: unreliable start\n') return if block['matches'][-1] < 0.5: sys.stderr.write('discard: unreliable end\n') return if np.mean(block['matches']) < 0.7: sys.stderr.write('discard: unreliable mean\n') return if len(block['matches']) < 5: sys.stderr.write('discard: too few words\n') return l = block['audioend'] - block['audiostart'] if l < 12: sys.stderr.write('discard: too short\n') return if l > 30: sys.stderr.write('discard: too long\n') return sys.stderr.write('accept\n') print("%f\t%f\t%s\t%s" % (block['audiostart'], block['audioend'], block['startlineno'], block['endlineno'])) start_line = 0 start_time = 0 spliti = 0 block = init_block() last_split = 0 for line in sys.stdin: (startstr, endstr, slenstr, predword, goldword, goldlinenostr, matchstr) = line.rstrip().split("\t") if block['startlineno'] == None: block['startlineno'] = goldlinenostr block['audiostart'] = last_split match = float(matchstr) block['matches'].append(match) block['endlineno'] = goldlinenostr end = float(endstr) if spliti < len(splits) and end > splits[spliti]: block['audioend'] = end last_split = splits[spliti] process_block(block) block = init_block() spliti = spliti + 1 process_block(block)
[ "jan@sixtease.net" ]
jan@sixtease.net
27e3a773e1f3b1c7193ce9a831b0b54a38653ad7
cf5f24e5a32f8cafe90d4253d727b1c0457da6a4
/algorithm/BOJ_1629.py
11a30af639ff558eb56b49660735d2acd32acf3e
[]
no_license
seoljeongwoo/learn
537659ca942875f6846646c2e21e1e9f2e5b811e
5b423e475c8f2bc47cb6dee09b8961d83ab08568
refs/heads/main
2023-05-04T18:07:27.592058
2021-05-05T17:32:50
2021-05-05T17:32:50
324,725,000
0
1
null
null
null
null
UTF-8
Python
false
false
290
py
# import sys # input = sys.stdin.readline # A,B,C=map(int,input().split()) # def solve(a,b): # if b==1: return a # ret = solve(a,b//2)%C # ret = (ret*ret)%C # if b%2==1: ret = (ret*a)%C # return ret # print(solve(A,B)%C) print(pow(*map(int,input().split())))
[ "noreply@github.com" ]
noreply@github.com
f1648f00e3328ab6336ed5fbb02a0739fce86d09
c545e20c20d4f3a9e580b543e0ff79555972ab94
/prepare_seq2seq.py
fb95072fec543d356525e3bf627763e0b5eab1e9
[]
no_license
StNiki/MT_CW3
d6cd6f33865653413e4bc1c7c992cfb96e28d894
b3c4f276396e8539f7473390a1cbf84770d28780
refs/heads/master
2021-01-21T14:43:24.967123
2017-06-26T21:34:06
2017-06-26T21:34:06
95,326,694
0
0
null
null
null
null
UTF-8
Python
false
false
5,261
py
# coding: utf-8 # ## Prepare parallel corpus # # **Based on TensorFlow code: https://github.com/tensorflow/models/blob/master/tutorials/rnn/translate/data_utils.py** # In[ ]: import os import re import pickle from tqdm import tqdm import sys # In[ ]: from nmt_config import * # In[ ]: data_fname = {"en": os.path.join(data_dir, "text_all.en"), "fr": os.path.join(data_dir, "text_all.fr")} # In[ ]: # Regular expressions used to tokenize. _WORD_SPLIT = re.compile(b"([.,!?\"':~;)(])") _DIGIT_RE = re.compile(br"\d") # In[ ]: def basic_tokenizer(sentence): """Very basic tokenizer: split the sentence into a list of tokens.""" words = [] for space_separated_fragment in sentence.strip().split(): words.extend(_WORD_SPLIT.sub(b"", w) for w in _WORD_SPLIT.split(space_separated_fragment)) return [w.lower() for w in words if w] # In[ ]: def extract_k_lines(fr_fname, en_fname, k): num_lines = 0 with open(data_fname["fr"],"rb") as f_fr, open(data_fname["en"],"rb") as f_en: with open(fr_fname,"wb") as out_fr, open(en_fname,"wb") as out_en: for i, (line_fr, line_en) in enumerate(zip(f_fr, f_en)): if num_lines >= k: break words_fr = basic_tokenizer(line_fr) words_en = basic_tokenizer(line_en) if len(words_fr) > 0 and len(words_en) > 0: # write to tokens file out_fr.write(b" ".join(words_fr) + b"\n") out_en.write(b" ".join(words_en) + b"\n") num_lines += 1 print("Total lines={0:d}, valid lines={1:d}".format(i, num_lines)) print("finished writing {0:s} and {1:s}".format(fr_fname, en_fname)) # In[ ]: def create_vocab(text_fname, num_train, max_vocabulary_size, freq_thresh): vocab = {} w2i = {} i2w = {} with open(text_fname,"rb") as in_f: for i, line in enumerate(in_f): if i >= num_train: break words = line.strip().split() for w in words: word = _DIGIT_RE.sub(b"0", w) word = _WORD_SPLIT.sub(b"", w) if word in vocab: vocab[word] += 1 else: vocab[word] = 1 print("vocab length before: {0:d}".format(len(vocab))) vocab = {k:vocab[k] for k in vocab if vocab[k] > freq_thresh} print("vocab length after: {0:d}".format(len(vocab))) vocab_list = START_VOCAB + sorted(vocab, key=vocab.get, reverse=True) print("Finished generating vocabulary") if len(vocab_list) > max_vocabulary_size: print("Vocab size={0:d}, trimmed to max={1:d}".format(len(vocab_list), max_vocabulary_size)) vocab_list = vocab_list[:max_vocabulary_size] else: print("Vocab size={0:d}".format(len(vocab_list))) for i, w in enumerate(vocab_list): w2i[w] = i i2w[i] = w print("finished vocab processing for {0:s}".format(text_fname)) for k in vocab: if vocab[k] <= freq_thresh: print("Ahaaaaa!!!", k, vocab[k]) return vocab, w2i, i2w # In[ ]: def create_input_config(k, num_train=NUM_TRAINING_SENTENCES, freq_thresh=FREQ_THRESH): # Output file names if not os.path.exists(input_dir): os.makedirs(input_dir) en_name = os.path.join(input_dir, "text.en") fr_name = os.path.join(input_dir, "text.fr") en_tokens_name = os.path.join(input_dir, "tokens.en") fr_tokens_name = os.path.join(input_dir, "tokens.fr") vocab_path = os.path.join(input_dir, "vocab.dict") w2i_path = os.path.join(input_dir, "w2i.dict") i2w_path = os.path.join(input_dir, "i2w.dict") # extract k lines extract_k_lines(fr_name, en_name, k) # create vocabularies vocab = {"en":{}, "fr":{}} w2i = {"en":{}, "fr":{}} i2w = {"en":{}, "fr":{}} print("*"*50) print("en file") print("*"*50) vocab["en"], w2i["en"], i2w["en"] = create_vocab(en_name, num_train=NUM_TRAINING_SENTENCES, max_vocabulary_size=max_vocab_size["en"], freq_thresh=FREQ_THRESH) print("*"*50) print("fr file") print("*"*50) vocab["fr"], w2i["fr"], i2w["fr"] = create_vocab(fr_name, num_train=NUM_TRAINING_SENTENCES, max_vocabulary_size=max_vocab_size["fr"], freq_thresh=FREQ_THRESH) print("*"*50) pickle.dump(vocab, open(vocab_path, "wb")) pickle.dump(w2i, open(w2i_path, "wb")) pickle.dump(i2w, open(i2w_path, "wb")) print("finished creating input config for {0:d} lines".format(k)) # In[ ]: create_input_config(k=NUM_SENTENCES, num_train=NUM_TRAINING_SENTENCES, freq_thresh=FREQ_THRESH) # In[ ]:
[ "noreply@github.com" ]
noreply@github.com
09eb58fd8a9817910ab1f48df9c299357ccadf50
1dff039deadc84ee5d9c92f6bee99baae03fb0dd
/horizons/gui/style.py
fd601b65699f124417027f6556eaad13bc7f351b
[]
no_license
totycro/unknown-horizons-quadtree
bf9321a204c907fea877125b95519d8574c9543a
6112ee6961714f6b963652d4ec25e2823732be8d
refs/heads/master
2021-01-18T14:21:53.161268
2011-03-22T23:04:27
2011-03-22T23:04:27
851,912
1
0
null
null
null
null
UTF-8
Python
false
false
7,414
py
# ################################################### # Copyright (C) 2009 The Unknown Horizons Team # team@unknown-horizons.org # This file is part of Unknown Horizons. # # Unknown Horizons is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the # Free Software Foundation, Inc., # 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA # ################################################### from fife import fife STYLES= { 'default': { 'default' : { 'border_size': 2, 'margins': (0, 0), 'base_color' : fife.Color(40, 40, 40, 0), 'foreground_color' : fife.Color(255, 255, 255), 'background_color' : fife.Color(40, 40, 40, 255), 'selection_color' : fife.Color(80, 80, 80, 255), 'font' : 'libertine_small' }, 'Button' : { 'border_size': 1, 'margins' : (10, 5) }, 'CheckBox' : { 'border_size': 0, 'background_color' : fife.Color(0, 0, 0, 0) }, 'RadioButton' : { 'border_size': 0, 'background_color' : fife.Color(0, 0, 0, 0) }, 'Label' : { 'border_size': 0, 'background_color' : fife.Color(40, 40, 40, 0), 'font' : 'libertine_small' }, 'ClickLabel' : { 'border_size': 0, 'font' : 'libertine_small' }, 'ListBox' : { 'border_size': 0, 'font' : 'libertine_small' }, 'Window' : { 'border_size': 1, 'margins': (10, 10), 'titlebar_height' : 30, 'font' : 'libertine_large', 'base_color' : fife.Color(60, 60, 60) }, 'TextBox' : { 'font' : 'libertine_small' }, ('Container','HBox','VBox') : { 'opaque' : 0, 'border_size': 0 }, ('Icon', 'ImageButton', 'DropDown') : { 'border_size': 0 }, 'ScrollArea' : { 'border_size': 0, 'horizontal_scrollbar' : 1, 'base_color' : fife.Color(60, 60, 60), 'background_color' : fife.Color(60, 60, 60) }, #'TextField' : { # 'base_color' : fife.Color(60, 60, 60), # 'background_color' : fife.Color(0, 0, 0) #} 'Slider' : { 'base_color' : fife.Color(80,80,40,50), }, }, 'menu': { #Used in the main menu and game menu 'default' : { 'border_size': 0, 'margins': (0, 0), 'opaque': 0, 'base_color' : fife.Color(0, 0, 0, 0), 'foreground_color' : fife.Color(255, 255, 255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0, 0, 0, 0), 'font' : 'libertine_mainmenu' }, 'Button' : { 'border_size': 0, 'margins' : (10, 5) }, 'Label' : { 'border_size': 0, 'font' : 'libertine_mainmenu' } }, 'menu_black': { # style for build menu etc. 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(255,255,255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0,0,0,0), 'font' : 'libertine_small_black' }, 'Button' : { 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'margins': (0,0), 'font' : 'libertine_14_black' } }, 'resource_bar': { 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0, 0, 0, 0), 'foreground_color' : fife.Color(0, 0, 0, 0), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0, 0, 0, 0), 'font' : 'libertine_small_black' }, 'Button' : { 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'alpha':0, 'font' : 'libertine_small_black' } }, 'message_text': { 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(255,255,255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0,0,0,0), 'font' : 'libertine_small' }, 'Button' : { 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'margins': (0,0), 'font' : 'libertine_small' } }, 'city_info': { # style for city info 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(255,255,255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0,0,0,0), 'font' : 'libertine_large' }, 'Button' : { 'font' : 'libertine_18', 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'font' : 'libertine_18' }, 'TooltipLabel': { 'font' : 'libertine_18' } }, 'headline': { # style for headlines 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(255,255,255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0,0,0,0), 'font' : 'libertine_headline' }, 'Button' : { 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'font' : 'libertine_headline' } }, 'book': { # style for book widgets 'default' : { 'border_size': 0, 'margins': (0,0), 'font' : 'libertine_14_black', 'foreground_color' : fife.Color(80,80,40), }, 'Label' : { 'font' : 'libertine_14_black', }, 'CheckBox' : { 'selection_color' : fife.Color(255,255,255,200), 'background_color' : fife.Color(255,255,255,128), 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(80,80,40), }, 'DropDown' : { 'selection_color' : fife.Color(255,255,255,200), 'background_color' : fife.Color(255,255,255,128), 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(80,80,40), 'font' : 'libertine_14_black', }, 'Slider' : { 'base_color' : fife.Color(80,80,40,128), }, 'TextBox' : { 'font' : 'libertine_14_black', 'opaque': 0 }, 'ListBox' : { 'background_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(80,80,40), 'selection_color' : fife.Color(255,255,255,128), 'font' : 'libertine_14_black', }, 'ScrollArea' : { 'background_color' : fife.Color(255,255,255,64), 'foreground_color' : fife.Color(80,80,40), 'base_color' : fife.Color(0,0,0,0), 'font' : 'libertine_14_black', 'horizontal_scrollbar' : 0, }, 'HBox' : { 'font' : 'libertine_14_black', 'foreground_color' : fife.Color(80,80,40), 'opaque': 0 }, 'TextField' : { 'selection_color' : fife.Color(255,255,255), 'background_color' : fife.Color(255,255,255,64), 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(80,80,40), 'font' : 'libertine_14_black', } }, 'tooltip': { # style for tooltips 'default' : { 'border_size': 0, 'margins': (0,0), 'opaque': 0, 'base_color' : fife.Color(0,0,0,0), 'foreground_color' : fife.Color(255,255,255), 'background_color' : fife.Color(0, 0, 0, 0), 'selection_color' : fife.Color(0,0,0,0), 'font' : 'libertine_headline' }, 'Button' : { 'border_size': 0, 'margins' : (0,0) }, 'Label' : { 'font' : 'libertine_tooltip' } }, }
[ "totycro@unknown-horizons.org" ]
totycro@unknown-horizons.org
d894c9f24bd98116d77798510af27cc57005c85c
f70c7ed21258a865c702c551aa8e251b88ff8f1d
/galloCRM/venv/bin/pip3
98d0d33d109ce13e0b03692ddfa17295bb175409
[]
no_license
gallofb/apple_CRM
9318f6bebce9bf81507870d625c72b47a16a46b8
b8c11a3ebced320af897d23f0954b93da9006a4c
refs/heads/master
2020-04-23T16:54:50.712398
2019-04-07T10:18:08
2019-04-07T10:18:08
171,314,207
0
0
null
null
null
null
UTF-8
Python
false
false
242
#!/home/gallo/apple/galloCRM/venv/bin/python # -*- coding: utf-8 -*- import re import sys from pip._internal import main if __name__ == '__main__': sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0]) sys.exit(main())
[ "854591086@qq.com" ]
854591086@qq.com
0eabf1e2f017b72b5cf73e83020af51b9ad7596d
d90e21b3250d5b3441465e71ad3e773e78058707
/rnd/HaskellRSLCompiler/test/parse/test.py
bef9a72ceb82bbb48832da89c306ea29b20a4752
[ "BSD-3-Clause", "LicenseRef-scancode-unknown-license-reference", "BSD-2-Clause" ]
permissive
staticagent/lucille
4ecf78ccede3efa9cd305d960c4172d1c2074065
ff81b332ae78181dbbdc1ec3c3b0f59992e7c0fa
refs/heads/master
2021-08-26T03:43:00.239208
2009-07-02T14:48:55
2009-07-02T14:48:55
null
0
0
null
null
null
null
UTF-8
Python
false
false
863
py
#!/usr/bin/env python import os, sys import subprocess import re import glob errlog = [] def run(f): cmd = "../../lslc" p = subprocess.Popen([cmd, f], stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True) outs = [l for l in p.stdout] errs = [l for l in p.stderr] errline = re.compile("TODO") failed = False for l in errs: if errline.search(l): failed = True if failed: print "[FAIL] ", f errlog.append("==== [" + f + "] ====") for l in errs: errlog.append(l[:-1]) errlog.append("=====================") errlog.append("\n") else: print "[OK ] ", f def main(): for f in glob.glob("*.sl"): run(f) f = open("errlog.log", "w") for l in errlog: print >>f, l if __name__ == '__main__': main()
[ "syoyo@lucillerender.org" ]
syoyo@lucillerender.org
7e70251ae9261b6cc83c7ebf3233459f5515f267
6fa701cdaa0d83caa0d3cbffe39b40e54bf3d386
/google/cloud/oslogin/v1beta/oslogin-v1beta-py/google/cloud/oslogin_v1beta/services/os_login_service/transports/grpc_asyncio.py
1c6fbed331f6c7d8d3fb9b348ed3be8e16af48ff
[ "Apache-2.0" ]
permissive
oltoco/googleapis-gen
bf40cfad61b4217aca07068bd4922a86e3bbd2d5
00ca50bdde80906d6f62314ef4f7630b8cdb6e15
refs/heads/master
2023-07-17T22:11:47.848185
2021-08-29T20:39:47
2021-08-29T20:39:47
null
0
0
null
null
null
null
UTF-8
Python
false
false
18,112
py
# -*- coding: utf-8 -*- # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # import warnings from typing import Awaitable, Callable, Dict, Optional, Sequence, Tuple, Union from google.api_core import gapic_v1 # type: ignore from google.api_core import grpc_helpers_async # type: ignore from google.auth import credentials as ga_credentials # type: ignore from google.auth.transport.grpc import SslCredentials # type: ignore import packaging.version import grpc # type: ignore from grpc.experimental import aio # type: ignore from google.cloud.oslogin.common import common_pb2 # type: ignore from google.cloud.oslogin_v1beta.types import oslogin from google.protobuf import empty_pb2 # type: ignore from .base import OsLoginServiceTransport, DEFAULT_CLIENT_INFO from .grpc import OsLoginServiceGrpcTransport class OsLoginServiceGrpcAsyncIOTransport(OsLoginServiceTransport): """gRPC AsyncIO backend transport for OsLoginService. Cloud OS Login API The Cloud OS Login API allows you to manage users and their associated SSH public keys for logging into virtual machines on Google Cloud Platform. This class defines the same methods as the primary client, so the primary client can load the underlying transport implementation and call it. It sends protocol buffers over the wire using gRPC (which is built on top of HTTP/2); the ``grpcio`` package must be installed. """ _grpc_channel: aio.Channel _stubs: Dict[str, Callable] = {} @classmethod def create_channel(cls, host: str = 'oslogin.googleapis.com', credentials: ga_credentials.Credentials = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, quota_project_id: Optional[str] = None, **kwargs) -> aio.Channel: """Create and return a gRPC AsyncIO channel object. Args: host (Optional[str]): The host for the channel to use. credentials (Optional[~.Credentials]): The authorization credentials to attach to requests. These credentials identify this application to the service. If none are specified, the client will attempt to ascertain the credentials from the environment. credentials_file (Optional[str]): A file with credentials that can be loaded with :func:`google.auth.load_credentials_from_file`. This argument is ignored if ``channel`` is provided. scopes (Optional[Sequence[str]]): A optional list of scopes needed for this service. These are only used when credentials are not specified and are passed to :func:`google.auth.default`. quota_project_id (Optional[str]): An optional project to use for billing and quota. kwargs (Optional[dict]): Keyword arguments, which are passed to the channel creation. Returns: aio.Channel: A gRPC AsyncIO channel object. """ return grpc_helpers_async.create_channel( host, credentials=credentials, credentials_file=credentials_file, quota_project_id=quota_project_id, default_scopes=cls.AUTH_SCOPES, scopes=scopes, default_host=cls.DEFAULT_HOST, **kwargs ) def __init__(self, *, host: str = 'oslogin.googleapis.com', credentials: ga_credentials.Credentials = None, credentials_file: Optional[str] = None, scopes: Optional[Sequence[str]] = None, channel: aio.Channel = None, api_mtls_endpoint: str = None, client_cert_source: Callable[[], Tuple[bytes, bytes]] = None, ssl_channel_credentials: grpc.ChannelCredentials = None, client_cert_source_for_mtls: Callable[[], Tuple[bytes, bytes]] = None, quota_project_id=None, client_info: gapic_v1.client_info.ClientInfo = DEFAULT_CLIENT_INFO, always_use_jwt_access: Optional[bool] = False, ) -> None: """Instantiate the transport. Args: host (Optional[str]): The hostname to connect to. credentials (Optional[google.auth.credentials.Credentials]): The authorization credentials to attach to requests. These credentials identify the application to the service; if none are specified, the client will attempt to ascertain the credentials from the environment. This argument is ignored if ``channel`` is provided. credentials_file (Optional[str]): A file with credentials that can be loaded with :func:`google.auth.load_credentials_from_file`. This argument is ignored if ``channel`` is provided. scopes (Optional[Sequence[str]]): A optional list of scopes needed for this service. These are only used when credentials are not specified and are passed to :func:`google.auth.default`. channel (Optional[aio.Channel]): A ``Channel`` instance through which to make calls. api_mtls_endpoint (Optional[str]): Deprecated. The mutual TLS endpoint. If provided, it overrides the ``host`` argument and tries to create a mutual TLS channel with client SSL credentials from ``client_cert_source`` or applicatin default SSL credentials. client_cert_source (Optional[Callable[[], Tuple[bytes, bytes]]]): Deprecated. A callback to provide client SSL certificate bytes and private key bytes, both in PEM format. It is ignored if ``api_mtls_endpoint`` is None. ssl_channel_credentials (grpc.ChannelCredentials): SSL credentials for grpc channel. It is ignored if ``channel`` is provided. client_cert_source_for_mtls (Optional[Callable[[], Tuple[bytes, bytes]]]): A callback to provide client certificate bytes and private key bytes, both in PEM format. It is used to configure mutual TLS channel. It is ignored if ``channel`` or ``ssl_channel_credentials`` is provided. quota_project_id (Optional[str]): An optional project to use for billing and quota. client_info (google.api_core.gapic_v1.client_info.ClientInfo): The client info used to send a user-agent string along with API requests. If ``None``, then default info will be used. Generally, you only need to set this if you're developing your own client library. always_use_jwt_access (Optional[bool]): Whether self signed JWT should be used for service account credentials. Raises: google.auth.exceptions.MutualTlsChannelError: If mutual TLS transport creation failed for any reason. google.api_core.exceptions.DuplicateCredentialArgs: If both ``credentials`` and ``credentials_file`` are passed. """ self._grpc_channel = None self._ssl_channel_credentials = ssl_channel_credentials self._stubs: Dict[str, Callable] = {} if api_mtls_endpoint: warnings.warn("api_mtls_endpoint is deprecated", DeprecationWarning) if client_cert_source: warnings.warn("client_cert_source is deprecated", DeprecationWarning) if channel: # Ignore credentials if a channel was passed. credentials = False # If a channel was explicitly provided, set it. self._grpc_channel = channel self._ssl_channel_credentials = None else: if api_mtls_endpoint: host = api_mtls_endpoint # Create SSL credentials with client_cert_source or application # default SSL credentials. if client_cert_source: cert, key = client_cert_source() self._ssl_channel_credentials = grpc.ssl_channel_credentials( certificate_chain=cert, private_key=key ) else: self._ssl_channel_credentials = SslCredentials().ssl_credentials else: if client_cert_source_for_mtls and not ssl_channel_credentials: cert, key = client_cert_source_for_mtls() self._ssl_channel_credentials = grpc.ssl_channel_credentials( certificate_chain=cert, private_key=key ) # The base transport sets the host, credentials and scopes super().__init__( host=host, credentials=credentials, credentials_file=credentials_file, scopes=scopes, quota_project_id=quota_project_id, client_info=client_info, always_use_jwt_access=always_use_jwt_access, ) if not self._grpc_channel: self._grpc_channel = type(self).create_channel( self._host, credentials=self._credentials, credentials_file=credentials_file, scopes=self._scopes, ssl_credentials=self._ssl_channel_credentials, quota_project_id=quota_project_id, options=[ ("grpc.max_send_message_length", -1), ("grpc.max_receive_message_length", -1), ], ) # Wrap messages. This must be done after self._grpc_channel exists self._prep_wrapped_messages(client_info) @property def grpc_channel(self) -> aio.Channel: """Create the channel designed to connect to this service. This property caches on the instance; repeated calls return the same channel. """ # Return the channel from cache. return self._grpc_channel @property def delete_posix_account(self) -> Callable[ [oslogin.DeletePosixAccountRequest], Awaitable[empty_pb2.Empty]]: r"""Return a callable for the delete posix account method over gRPC. Deletes a POSIX account. Returns: Callable[[~.DeletePosixAccountRequest], Awaitable[~.Empty]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'delete_posix_account' not in self._stubs: self._stubs['delete_posix_account'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/DeletePosixAccount', request_serializer=oslogin.DeletePosixAccountRequest.serialize, response_deserializer=empty_pb2.Empty.FromString, ) return self._stubs['delete_posix_account'] @property def delete_ssh_public_key(self) -> Callable[ [oslogin.DeleteSshPublicKeyRequest], Awaitable[empty_pb2.Empty]]: r"""Return a callable for the delete ssh public key method over gRPC. Deletes an SSH public key. Returns: Callable[[~.DeleteSshPublicKeyRequest], Awaitable[~.Empty]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'delete_ssh_public_key' not in self._stubs: self._stubs['delete_ssh_public_key'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/DeleteSshPublicKey', request_serializer=oslogin.DeleteSshPublicKeyRequest.serialize, response_deserializer=empty_pb2.Empty.FromString, ) return self._stubs['delete_ssh_public_key'] @property def get_login_profile(self) -> Callable[ [oslogin.GetLoginProfileRequest], Awaitable[oslogin.LoginProfile]]: r"""Return a callable for the get login profile method over gRPC. Retrieves the profile information used for logging in to a virtual machine on Google Compute Engine. Returns: Callable[[~.GetLoginProfileRequest], Awaitable[~.LoginProfile]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'get_login_profile' not in self._stubs: self._stubs['get_login_profile'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/GetLoginProfile', request_serializer=oslogin.GetLoginProfileRequest.serialize, response_deserializer=oslogin.LoginProfile.deserialize, ) return self._stubs['get_login_profile'] @property def get_ssh_public_key(self) -> Callable[ [oslogin.GetSshPublicKeyRequest], Awaitable[common_pb2.SshPublicKey]]: r"""Return a callable for the get ssh public key method over gRPC. Retrieves an SSH public key. Returns: Callable[[~.GetSshPublicKeyRequest], Awaitable[~.SshPublicKey]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'get_ssh_public_key' not in self._stubs: self._stubs['get_ssh_public_key'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/GetSshPublicKey', request_serializer=oslogin.GetSshPublicKeyRequest.serialize, response_deserializer=common_pb2.SshPublicKey.FromString, ) return self._stubs['get_ssh_public_key'] @property def import_ssh_public_key(self) -> Callable[ [oslogin.ImportSshPublicKeyRequest], Awaitable[oslogin.ImportSshPublicKeyResponse]]: r"""Return a callable for the import ssh public key method over gRPC. Adds an SSH public key and returns the profile information. Default POSIX account information is set when no username and UID exist as part of the login profile. Returns: Callable[[~.ImportSshPublicKeyRequest], Awaitable[~.ImportSshPublicKeyResponse]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'import_ssh_public_key' not in self._stubs: self._stubs['import_ssh_public_key'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/ImportSshPublicKey', request_serializer=oslogin.ImportSshPublicKeyRequest.serialize, response_deserializer=oslogin.ImportSshPublicKeyResponse.deserialize, ) return self._stubs['import_ssh_public_key'] @property def update_ssh_public_key(self) -> Callable[ [oslogin.UpdateSshPublicKeyRequest], Awaitable[common_pb2.SshPublicKey]]: r"""Return a callable for the update ssh public key method over gRPC. Updates an SSH public key and returns the profile information. This method supports patch semantics. Returns: Callable[[~.UpdateSshPublicKeyRequest], Awaitable[~.SshPublicKey]]: A function that, when called, will call the underlying RPC on the server. """ # Generate a "stub function" on-the-fly which will actually make # the request. # gRPC handles serialization and deserialization, so we just need # to pass in the functions for each. if 'update_ssh_public_key' not in self._stubs: self._stubs['update_ssh_public_key'] = self.grpc_channel.unary_unary( '/google.cloud.oslogin.v1beta.OsLoginService/UpdateSshPublicKey', request_serializer=oslogin.UpdateSshPublicKeyRequest.serialize, response_deserializer=common_pb2.SshPublicKey.FromString, ) return self._stubs['update_ssh_public_key'] __all__ = ( 'OsLoginServiceGrpcAsyncIOTransport', )
[ "bazel-bot-development[bot]@users.noreply.github.com" ]
bazel-bot-development[bot]@users.noreply.github.com
4dd6d2950da0380ec4d98105143ddc84def25d6e
a878fbb811d5f49b0524b4a03710b96d7f6f88b6
/old/data_suff/utils.py
5d2180887b52f252c95ad5e3212b82a5e3be003b
[ "MIT" ]
permissive
alonshpigler/CovidScreening
caad134a98aaafe88dbd2cd5d24802ef833f5200
f057c6197db6550693ff6b606b3d8c264508aba5
refs/heads/main
2023-07-02T17:50:09.292805
2021-08-15T10:45:47
2021-08-15T10:45:47
361,657,171
0
0
null
null
null
null
UTF-8
Python
false
false
1,286
py
import os # from skimage.io import imread import pandas as pd from util import load_csv # # filename = 'D:\\RxRx19a-images.zip' # with ZipFile(filename) as archive: # for entry in archive.infolist(): # with archive.open(entry) as file: # img = Image.open(file) # print(img.size, img.mode, len(img.getdata())) # DEFAULT_BASE_PATH = 'C:/Covid-Screening/data_layer/raw_data' DEFAULT_METADATA_BASE_PATH = os.path.join(DEFAULT_BASE_PATH, 'metadata') DEFAULT_IMAGES_BASE_PATH = os.path.join(DEFAULT_BASE_PATH, 'images') DEFAULT_CHANNELS = (1, 2, 3, 4, 5) def _load_dataset(base_path, dataset, include_controls=True): df = load_csv(os.path.join(base_path, dataset + '.csv')) if include_controls: controls = load_csv( os.path.join(base_path, dataset + '_controls.csv')) df['well_type'] = 'treatment' df = pd.concat([controls, df], sort=True) df['cell_type'] = df.experiment.str.split("-").apply(lambda a: a[0]) df['dataset'] = dataset dfs = [] for site in (1, 2): df = df.copy() df['site'] = site dfs.append(df) res = pd.concat(dfs).sort_values( by=['id_code', 'site']).set_index('id_code') return res _load_dataset(DEFAULT_BASE_PATH,'metadata',False)
[ "alonshp@post.bgu.ac.il" ]
alonshp@post.bgu.ac.il
f23530b0fcab203fccb0a43b9d3560015edbb1df
07504838d12c6328da093dce3726e8ed096cecdb
/pylon/resources/properties/safExtCnfg.py
f32c8cf1cee106166c91aab4c960446be0295d8e
[]
no_license
lcoppa/fiat-lux
9caaa7f3105e692a149fdd384ec590676f06bf00
7c166bcc08768da67c241078b397570de159e240
refs/heads/master
2020-04-04T02:47:19.917668
2013-10-10T10:22:51
2013-10-10T10:22:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
2,213
py
"""safExtCnfg standard property type, originally defined in resource file set standard 00:00:00:00:00:00:00:00-0.""" # Copyright (C) 2013 Echelon Corporation. All Rights Reserved. # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the "Software" to # deal in the Software without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, and/or # sell copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL # THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. # This file is generated from device resource files using an automated # database to source code conversion process. Grammar and punctuation within # the embedded documentation may not be correct, as this data is gathered and # combined from several sources. The machine-generated code may not meet # compliance with PEP-8 and PEP-257 recommendations at all times. # Generated at 23-Sep-2013 09:14. import pylon.resources.base from pylon.resources.standard import standard class safExtCnfg(pylon.resources.base.Inheriting): """safExtCnfg standard property type. Safety mode. Mode that a device has to be brought to when a safety external request state is pending.""" def __init__(self): super().__init__( ) self._original_name = 'SCPTsafExtCnfg' self._property_scope, self._property_key = 0, 257 self._definition = standard.add(self) if __name__ == '__main__': # unit test code. item = safExtCnfg() pass
[ "lcoppa@rocketmail.com" ]
lcoppa@rocketmail.com
0cd738fe1ed690b6bda48d0d7ac79cc298310385
e42174d0b04e2ff33955a1d49297d8deaa6a1cfe
/Main2.py
75e16c0fa236f8c6b7ec947ccca6ba715002a0bc
[]
no_license
Fudan-iGEM/2021-software-parse
ead1669940aff560ed74ccc54e6d488657792714
e604a1cedd49f7b50a7d5bdcab2a5f4475d0bcf4
refs/heads/main
2023-08-01T08:02:13.727207
2021-09-18T14:45:13
2021-09-18T14:45:13
419,995,424
1
0
null
null
null
null
UTF-8
Python
false
false
14,400
py
import requests from bs4 import BeautifulSoup import time import re import multiprocessing from multiprocessing import Pool from openpyxl import workbook from openpyxl import load_workbook #from lxml import etree from selenium import webdriver from selenium.webdriver.common.action_chains import ActionChains from selenium.webdriver.support import expected_conditions as EC from selenium.webdriver.common.by import By from selenium.webdriver.support.wait import WebDriverWait #全篇的sleep函数可以优化 #打不开的情况需要try函数优化 #整体上,以年为记的part为全局变量 class Part: # part_num(BBa.), part_name(CAP), part_id(内部代码), part_type(com/pro...),star(包含队伍使用和独特的标记) def __init__(self, part_num, part_name, part_id, part_url, short_desc, part_type, team, year, sequence, contents, stars, assemble_std, linking_parts, parts_used, using_parts, len): self.part_num = part_num self.part_name = part_name self.part_id = part_id self.part_url = part_url self.short_desc = short_desc self.year = year self.sequence = sequence #stars 需要再细化 self.stars = stars self.assemble_std = assemble_std self.contents = contents # self.linking_parts = linking_parts # how to get self.parts_used = parts_used self.using_parts = using_parts self.len = len self.part_type = part_type self.team = team def print_parts(self): print(f"part_num = {self.part_num}") print(f"part_name = {self.part_name}") print(f"part_id = {self.part_id}") print(f"part_url = {self.part_url}") print(f"part_type = {self.part_type}") print(f"part_team = {self.team}") print(f"part_year = {self.year}") print(f"part_sequence = {self.sequence}") print(f"part_stars = {self.stars}") print(f"part_desc = {self.short_desc}") print(f"part_assemble_std = {self.assemble_std}") print(f"contents"f" = {self.contents}") print(f"parts_used = {self.parts_used}") print(f"using_parts = {self.using_parts}") print(f"len = {self.len}") print("------------------------------") all_team_with_urls = [] whole_Parts = [] def inter(): # 需要标注输入格式,届时可以通过前端重写 print("Which year would you want to scan for? Input 'years'") return 0 #已完成,返回的是某一年的所有队伍的url def web_analysis_and_get_team_lists(year): # year 可变 # 此处的地址可能需要更改 #desktop地址: ‘D:\chromedriver.exe’ #laptop地址:'C:\Python x64\Python\chromedriver.exe' print(f"---Start getting team lists in {year}---") driver = webdriver.Chrome('C:\\Users\GhoST\AppData\Local\Google\Chrome\Application\ChromeDriver.exe') front_url = "https://igem.org/Team_Parts?year=" url = front_url + year #此处未检查! while 1: try: driver.get(url) WebDriverWait(driver, 60, 1).until( EC.presence_of_element_located((By.XPATH, '//*[@id="topBanner"]/a/img')), message='') break except: print("刷新") driver.refresh() pass time.sleep(1) one_team_with_url = [] the_list = driver.find_elements_by_xpath('/html/body/div/div[3]/div/div/div/div[4]/table/tbody/tr/td/div/a') for item in the_list: one_team_with_url = [year, str(item.text), str(item.get_attribute('href'))] all_team_with_urls.append(one_team_with_url) print(f"---Ending getting team lists in {year}---") driver.close() return all_team_with_urls #未完成 def set_star_database(): return #输入某一年所有队伍的url,输出这一年所有part的基础信息(全局变量中) def get_parts_urls(all_team_with_urls): print("---Start getting parts urls and basic info--- ") for a_team in all_team_with_urls: year = a_team[0] team = a_team[1] url = a_team[2] # desktop地址: ‘D:\chromedriver.exe’ # laptop地址:'C:\Python x64\Python\chromedriver.exe' driver = webdriver.Chrome('C:\\Users\GhoST\AppData\Local\Google\Chrome\Application\ChromeDriver.exe') while 1: try: driver.get(url) WebDriverWait(driver, 5, 1).until(EC.presence_of_element_located((By.XPATH, '//*[@id="new_menubar"]/ul/li[1]/div[1]')), message = '') break except: print("刷新") driver.refresh() pass time.sleep(1) #先将基础属性放在列表里 part_num_list = [] part_numurl_list = [] part_type_list = [] part_desc = [] part_designer = [] part_len = [] #得到第一张表的数据(favored) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[1]/tbody/tr/td[3]/a'): part_num_list.append(str(item.text)) part_numurl_list.append(item.get_attribute('href')) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[1]/tbody/tr/td[4]'): part_type_list.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[1]/tbody/tr/td[5]'): part_desc.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[1]/tbody/tr/td[6]'): part_designer.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[1]/tbody/tr/td[7]'): part_len.append(str(item.text)) #为第第一张表(favored)创建类 #star第一个1代表favor for i in range(0, len(part_num_list)): new_part = Part(part_num_list[i], '', '', part_numurl_list[i], part_desc[i], part_type_list[i], team, year, '', '', '1', '', [], [], [], part_len[i]) whole_Parts.append(new_part) part_num_list = [] part_numurl_list = [] part_type_list = [] part_desc = [] part_designer = [] part_len = [] #得到第二张表的数据(NOT favored) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[2]/tbody/tr/td[3]/a'): part_num_list.append(str(item.text)) part_numurl_list.append(item.get_attribute('href')) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[2]/tbody/tr/td[4]'): part_type_list.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[2]/tbody/tr/td[5]'): part_desc.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[2]/tbody/tr/td[6]'): part_designer.append(str(item.text)) for item in driver.find_elements_by_xpath('/html/body/div/div[4]/div/div/table[2]/tbody/tr/td[7]'): part_len.append(str(item.text)) # 为第第一张表(favored)创建类 # star第一个1代表favor for i in range(0, len(part_num_list)): new_part = Part(part_num_list[i], '', '', part_numurl_list[i], part_desc[i], part_type_list[i], team, year, '', '', '0', '', [], [], [], part_len[i]) whole_Parts.append(new_part) print("---End getting parts urls and basic info--- ") driver.close() return # 未完成,从全局PART中,开始进行操作 def get_parts_details(): for a_part in whole_Parts: print(f"---Start getting details of {a_part.part_num}---") url = a_part.part_url driver = webdriver.Chrome('C:\\Users\GhoST\AppData\Local\Google\Chrome\Application\ChromeDriver.exe') i = 0 gotten = False while 1: try: driver.get(url) WebDriverWait(driver, 10, 1).until(EC.presence_of_element_located((By.XPATH, '//*[@id="new_menubar"]/ul/li[1]/div[1]')), message = '') gotten = True break except: i= i + 1 if i > 5: break print("刷新") driver.refresh() pass time.sleep(1) #以上打开了part的主网页界面 if not gotten : #new_part = Part(part_num_list[i], '', '', part_numurl_list[i], part_desc[i], part_type_list[i], team, year, '', '', '0', '', [], [], [], part_len[i]) continue #------------------------------------------- get_using_parts_and_other_info(driver, a_part) get_assemble_std(driver, a_part) get_used_parts(driver, a_part) #GET_SEQUENCE 自带关闭整个窗口的作用,所以所有数据获取请在这句之前玩完成 get_sequence(driver, a_part) # ------------------------------------------- print(f"---End getting details of {a_part.part_num}---") store_parts() print(f"---Details of parts in {a_part.year} are saved---") return #已完成,used代表使用了该part的part,需要额外打开页面 def get_used_parts(driver, a_part): try: item = driver.find_elements_by_xpath('//*[@id="part_status_wrapper"]/div[4]/a') url = str(item[0].get_attribute('href')) except: a_part.parts_used = 'None' return while 1: try: driver.get(url) WebDriverWait(driver, 10, 1).until( EC.presence_of_element_located((By.XPATH,'/html/body')), message='') break except: print("刷新") driver.refresh() pass time.sleep(1) used_parts = [] list = driver.find_elements_by_class_name('noul_link.part_link') for item in list: used_parts.append(str(item.text)) if len(used_parts) == 0: used_parts.append('None') a_part.parts_used = used_parts driver.back() return #已完成 def get_assemble_std(driver, a_part): assemble_lists = [] for item in driver.find_elements_by_xpath('//*[@id="assembly_compatibility"]/div/ul/li'): if str(item.get_attribute("class")) == "boxctrl box_green" : assemble_lists.append('1') else: assemble_lists.append('0') #assemble_lists.append(str(item.get_attribute("class"))) a_part.assemble_std = assemble_lists return #这一部分写part主页面内的所有内容。using代表该part的组成part,不需要额外打开页面加载; #已完成,不关闭窗口 def get_using_parts_and_other_info(driver, a_part): if a_part.part_type != 'Composite': a_part.using_parts = ['self'] else: using_parts_list = [] for item in driver.find_elements_by_xpath('//*[@id="seq_features_div"]/div[1]/div[4]/div/div[2]'): using_parts_list.append(str(item.text)) #以下确认了编号的统一 for i in range(0, len(using_parts_list)): if 'BBa' in using_parts_list[i]: continue else: using_parts_list[i] = 'BBa_'+ using_parts_list[i] a_part.using_parts = using_parts_list return #自带关闭,已完成 def get_sequence(driver, a_part): sequence_entrance = driver.find_elements_by_xpath('//*[@id="seq_features_div"]/div[1]/div[1]/span[5]') #webdriver.ActionChains(driver).move_to_element(sequence_entrance[0]).click(sequence_entrance[0]).perform().find_elements_by_xpath("/html/body/pre/text()") try: webdriver.ActionChains(driver).move_to_element(sequence_entrance[0]).click(sequence_entrance[0]).perform() except: print(f"{a_part.part_num} 没有序列或序列获取失败") return time.sleep(1) #切换窗口到新跳出的窗口 handles = driver.window_handles index_handle = driver.current_window_handle#备注:可能需要在操作前,先关闭其他浏览器窗口 for handle in handles: if handle != index_handle: driver.switch_to.window(handle) sequence = driver.find_elements_by_xpath("/html/body/pre")#备注:所有xpath出来都是list,记得切换为元素 a_part.sequence = str(sequence[0].text) driver.close() handle = driver.window_handles[0] driver.switch_to.window(handle) driver.close() return #下一个需要完成的,一年一存 def store_parts(): wb = workbook.Workbook() ws1 = wb.active ws1.append(['part_num', 'part_name', 'part_id', 'part_url', 'short_desc', 'part_type', 'team', 'year', 'sequence', 'contents', 'stars', 'assemble_std', 'parts_used', 'using_parts', 'len']) for a_part in whole_Parts: ws1.append([a_part.part_num, a_part.part_name, a_part.part_id, a_part.part_url, a_part.short_desc, \ a_part.part_type, a_part.team, a_part.year, a_part.sequence, a_part.contents, a_part.stars,\ ' '.join(a_part.assemble_std), ' '.join(a_part.parts_used), ' '.join(a_part.using_parts), a_part.len]) wb.save(f'D:\\{a_part.year}collection.xlsx') return def main(): #year = '2004' # for year in range(2020): #all_team_with_urls = web_analysis_and_get_team_lists(str(year)) #anothoer test_example: ['2020','teamB',' http://parts.igem.org/cgi/partsdb/pgroup.cgi?pgroup=iGEM2020&group=Fudan'] ''' all_team_with_urls = [['2019', 'teamA', 'http://parts.igem.org/cgi/partsdb/pgroup.cgi?pgroup=iGEM2020&group=GDSYZX']] get_parts_urls(all_team_with_urls) # 所有信息存在全局变量 whole_Parts 中 get_parts_details() # 所有信息存在全局变量 whole_Parts 中,并且一个一存/一年一存 return 0 ''' years = [2021] for year in years: all_team_with_urls = web_analysis_and_get_team_lists(str(year)) get_parts_urls(all_team_with_urls) #所有信息存在全局变量 whole_Parts 中 get_parts_details() #所有信息存在全局变量 whole_Parts 中,并且一年一存 whole_Parts = [] all_team_with_urls = [] return 0 main()
[ "TomGhostSmith@gmail.com" ]
TomGhostSmith@gmail.com
571fa44c13cd529ee4eb84aca37e4931096a4b98
d22dc2f7cbf5370c7afc381b27efffcdf8a7d00a
/prob05.py3
42e2e8b5719a19f62ba366bbaed5041b56a6acbd
[]
no_license
xorkevin/CodeWarsHP2016
13150154f969bb1765ecf4206c7f5106f93e8089
2732c3944e2431d43d574428b70b0e48a1e623ee
refs/heads/master
2021-01-10T01:21:43.580960
2016-03-05T20:11:12
2016-03-05T20:11:12
53,221,211
0
0
null
null
null
null
UTF-8
Python
false
false
299
py3
import sys f = sys.stdin.read() f = f.split('\n') f.pop(0) f.pop() for line in f: x, y = line.split(' ') x = int(x) y = list(enumerate(y)) y = list(filter(lambda tup: tup[0]%x != 0, y)) k = '' for tup in y: k += tup[1] print('' + str(k) + ' ' + str(len(k)))
[ "wangkevin448@gmail.com" ]
wangkevin448@gmail.com
003433cb893cff17a7ae9e5807ff49deed068997
0cc4eb3cb54f8394c127ace62d3108fdb5230c85
/.spack-env/view/lib/python3.7/site-packages/jedi/third_party/typeshed/stdlib/2/dircache.pyi
523b850bc3e93f867de75c9bef0100e3b6d22c54
[]
no_license
jacobmerson/spack-develop-env
5b2d76f58c0b64ae97c64f77a3c4d33a770c71c8
5fca20ca343b1a76f05fc635c87f94ed25417d94
refs/heads/master
2022-07-04T02:22:50.264727
2020-05-06T05:13:50
2020-05-06T05:13:50
261,657,112
0
0
null
null
null
null
UTF-8
Python
false
false
187
pyi
/lore/mersoj/spack/spack/opt/spack/linux-rhel7-x86_64/gcc-7.3.0/py-jedi-0.17.0-zugnvpgjfmuk5x4rfhhxlsknl2g226yt/lib/python3.7/site-packages/jedi/third_party/typeshed/stdlib/2/dircache.pyi
[ "mersoj@rpi.edu" ]
mersoj@rpi.edu
1ba87cd411f46c264b9fd8759ef716c3d9e27938
c06efd90533c51c2b29b7e92cd13723388de25ee
/actions/patchStorageV1beta1StorageClass.py
a57bbce258efa5ad9e6ef149ec1d897e8648932f
[]
no_license
ajohnstone/stackstorm-kubernetes
490e4a73daad3713d7c5b5b639d5f30ff1ab3e58
99ffad27f5947583a2ab1b56e80c06003d014c47
refs/heads/master
2021-01-11T23:29:49.642435
2016-12-07T13:20:34
2016-12-07T13:20:34
78,588,572
0
0
null
2017-01-11T00:48:59
2017-01-11T00:48:59
null
UTF-8
Python
false
false
746
py
from lib import k8s from st2actions.runners.pythonrunner import Action class patchStorageV1beta1StorageClass(Action): def run(self,body,name,config_override=None,pretty=None): myk8s = k8s.K8sClient(self.config) args = {} if body is not None: args['body'] = body else: return (False, "body is a required parameter") if name is not None: args['name'] = name else: return (False, "name is a required parameter") if config_override is not None: args['config_override'] = config_override if pretty is not None: args['pretty'] = pretty return (True, myk8s.runAction('patchStorageV1beta1StorageClass', **args))
[ "andy@impulsed.net" ]
andy@impulsed.net
a82e3b55c739f26cf5bb75daf2ce5b13f71cf467
7997c79e65a20e80ba76887181378ad110389121
/making_queries/apps.py
c21846fc2f0c4688682d053d2abe910725fe1e01
[]
no_license
caerang/django-practice
68b7940087fdab51dc2e4ab589e0d1bab3e469e6
cd9497053ec16f19c70daba03c0117d710a8b64b
refs/heads/master
2022-12-06T04:32:33.084302
2022-11-28T07:09:46
2022-11-28T07:09:46
80,974,634
0
0
null
null
null
null
UTF-8
Python
false
false
107
py
from django.apps import AppConfig class MakingQueriesConfig(AppConfig): name = 'making_queries'
[ "mylovercorea@gmail.com" ]
mylovercorea@gmail.com
5b954273bea26961fa06977aa411ddafbbcf44d9
a31bb4bb77fb3a40b00ac6b3e4cdaf3bb4fa54d7
/day1/while.py
576c054112ff82b57df49e5e7b9422d3faa9b507
[]
no_license
lin790292154/python_note
457069a5d0adc6991321f5eb4644e1c91c597fff
3b35a3087117aadc0b8b63a260ef55e5b20f5e96
refs/heads/master
2021-08-31T19:58:15.775229
2017-12-22T17:14:40
2017-12-22T17:14:40
110,909,371
0
0
null
null
null
null
UTF-8
Python
false
false
448
py
# write by kratos ''' count = 0 while True: print("count=",count) count = count + 1 ''' age_of_oldboy = 66 count = 0 while count < 3 : guess_age = int(input("what is oldboy age ?")) if age_of_oldboy == guess_age : print("you are right") break elif age_of_oldboy > guess_age: print("To small") else: print("To big") count += 1 else: print("you have tried too many time,fuck off")
[ "lin790292154@163.com" ]
lin790292154@163.com
d330066fb6ba0e836748a43a60059fe223936d8f
7e53ed2d6074a025fe960f72c21672cc23dcab14
/vt/tests/test_vt.py
c1d05f734a3b1993f80ddc0c57020fbbb90a49cb
[ "MIT" ]
permissive
kyokley/vittlify-cli
154410638b3a33640c01ab915dbf24d4e6afe13f
e3be7f3c7b0c00d59defe73af9aed0ec792800cc
refs/heads/master
2023-02-17T21:08:29.452548
2021-12-05T15:54:06
2021-12-05T15:54:06
58,974,128
0
0
MIT
2023-02-08T02:27:24
2016-05-16T23:09:51
Python
UTF-8
Python
false
false
41,993
py
import shlex import unittest import mock import pytest import requests from vt.utils import VittlifyError from vt.vt import ( Status, add, categories, complete, display_all_shopping_lists, display_item, display_shopping_list, display_shopping_list_categories, help, modify, move, run, show, term, ) class TestDisplayShoppingList(unittest.TestCase): def setUp(self): self.get_shopping_list_info_patcher = mock.patch('vt.vt.get_shopping_list_info') self.mock_get_shopping_list_info = self.get_shopping_list_info_patcher.start() self.get_shopping_list_items_patcher = mock.patch( 'vt.vt.get_shopping_list_items' ) self.mock_get_shopping_list_items = self.get_shopping_list_items_patcher.start() self.get_completed_patcher = mock.patch('vt.vt.get_completed') self.mock_get_completed = self.get_completed_patcher.start() self.get_all_shopping_list_items_patcher = mock.patch( 'vt.vt.get_all_shopping_list_items' ) self.mock_get_all_shopping_list_items = ( self.get_all_shopping_list_items_patcher.start() ) self.format_row_patcher = mock.patch('vt.vt.format_row') self.mock_format_row = self.format_row_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() test_shopping_list = {'name': 'test_list'} self.mock_get_shopping_list_info.return_value = test_shopping_list test_items = [ {'name': 'item1'}, {'name': 'item2'}, {'name': 'item3'}, ] self.mock_get_shopping_list_items.return_value = test_items self.mock_get_all_shopping_list_items.return_value = test_items self.mock_get_completed.return_value = test_items self.mock_format_row.side_effect = [ 'formatted_row_1', 'formatted_row_2', 'formatted_row_3', ] def tearDown(self): self.get_shopping_list_info_patcher.stop() self.get_shopping_list_items_patcher.stop() self.get_completed_patcher.stop() self.get_all_shopping_list_items_patcher.stop() self.format_row_patcher.stop() self.print_table_patcher.stop() def test_not_completed(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.NOT_COMPLETED) self.mock_get_shopping_list_info.assert_called_once_with(guid) self.mock_get_shopping_list_items.assert_called_once_with(guid) self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item2'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item3'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='test_list', quiet=False, ) def test_all(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.ALL) self.mock_get_shopping_list_info.assert_called_once_with(guid) self.mock_get_all_shopping_list_items.assert_called_once_with(guid) self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item2'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item3'}, {'name': 'test_list'}, include_category=False, include_comments=False, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='test_list', quiet=False, ) def test_completed(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.COMPLETED) self.assertFalse(self.mock_get_shopping_list_info.called) self.mock_get_completed.assert_called_once_with() self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, None, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item2'}, None, include_category=False, include_comments=False, no_wrap=False, ), mock.call( {'name': 'item3'}, None, include_category=False, include_comments=False, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='Recently Completed', quiet=False, ) def test_not_completed_extended(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.NOT_COMPLETED, extended=True) self.mock_get_shopping_list_info.assert_called_once_with(guid) self.mock_get_shopping_list_items.assert_called_once_with(guid) self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item2'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item3'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='test_list', quiet=False, ) def test_all_extended(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.ALL, extended=True) self.mock_get_shopping_list_info.assert_called_once_with(guid) self.mock_get_all_shopping_list_items.assert_called_once_with(guid) self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item2'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item3'}, {'name': 'test_list'}, include_category=False, include_comments=True, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='test_list', quiet=False, ) def test_completed_extended(self): guid = 'test_guid' display_shopping_list(guid=guid, mode=Status.COMPLETED, extended=True) self.assertFalse(self.mock_get_shopping_list_info.called) self.mock_get_completed.assert_called_once_with() self.mock_format_row.assert_has_calls( [ mock.call( {'name': 'item1'}, None, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item2'}, None, include_category=False, include_comments=True, no_wrap=False, ), mock.call( {'name': 'item3'}, None, include_category=False, include_comments=True, no_wrap=False, ), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='Recently Completed', quiet=False, ) class TestDisplayItem(unittest.TestCase): def setUp(self): self.get_item_patcher = mock.patch('vt.vt.get_item') self.mock_get_item = self.get_item_patcher.start() self.format_row_patcher = mock.patch('vt.vt.format_row') self.mock_format_row = self.format_row_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() self.test_guid = 'test_guid' def tearDown(self): self.get_item_patcher.stop() self.format_row_patcher.stop() self.print_table_patcher.stop() def test_(self): display_item(self.test_guid) self.mock_get_item.assert_called_once_with(self.test_guid) self.mock_format_row.assert_called_once_with( self.mock_get_item.return_value, None, include_comments=True, no_wrap=False ) self.mock_print_table.assert_called_once_with( [self.mock_format_row.return_value] ) class TestDisplayAllShoppingLists(unittest.TestCase): def setUp(self): self.get_all_shopping_lists_patcher = mock.patch('vt.vt.get_all_shopping_lists') self.mock_get_all_shopping_lists = self.get_all_shopping_lists_patcher.start() self.format_row_patcher = mock.patch('vt.vt.format_row') self.mock_format_row = self.format_row_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() self.mock_get_all_shopping_lists.return_value = [ {'name': 'list1'}, {'name': 'list2'}, {'name': 'list3'}, ] self.mock_format_row.side_effect = [ 'formatted_row_1', 'formatted_row_2', 'formatted_row_3', ] def tearDown(self): self.get_all_shopping_lists_patcher.stop() self.format_row_patcher.stop() def test_(self): display_all_shopping_lists() self.mock_get_all_shopping_lists.assert_called_once_with() self.mock_format_row.assert_has_calls( [ mock.call({'name': 'list1'}, None, no_wrap=False), mock.call({'name': 'list2'}, None, no_wrap=False), mock.call({'name': 'list3'}, None, no_wrap=False), ] ) self.mock_print_table.assert_called_once_with( ['formatted_row_1', 'formatted_row_2', 'formatted_row_3'], title='All Lists' ) class TestShowNoDefaultList(unittest.TestCase): def setUp(self): self.DEFAULT_LIST_patcher = mock.patch('vt.vt.DEFAULT_LIST', '') self.DEFAULT_LIST_patcher.start() self.display_shopping_list_patcher = mock.patch('vt.vt.display_shopping_list') self.mock_display_shopping_list = self.display_shopping_list_patcher.start() self.display_all_shopping_lists_patcher = mock.patch( 'vt.vt.display_all_shopping_lists' ) self.mock_display_all_shopping_lists = ( self.display_all_shopping_lists_patcher.start() ) self.display_item_patcher = mock.patch('vt.vt.display_item') self.mock_display_item = self.display_item_patcher.start() def tearDown(self): self.DEFAULT_LIST_patcher.stop() self.display_shopping_list_patcher.stop() self.display_all_shopping_lists_patcher.stop() self.display_item_patcher.stop() def test_list_empty_guid(self): args = shlex.split("list ''") self.assertRaises(IndexError, show, args) def test_list_no_guid(self): args = shlex.split("list") self.assertRaises(IndexError, show, args) def test_list_empty_guid_extended(self): args = shlex.split("list '' -e") self.assertRaises(IndexError, show, args) def test_list_no_guid_extended(self): args = shlex.split("list -e") self.assertRaises(IndexError, show, args) def test_list_no_extended(self): args = shlex.split("list test_guid") show(args) self.mock_display_shopping_list.assert_called_once_with(guid='test_guid') def test_list_extended(self): args = shlex.split("list test_guid -e") show(args) self.mock_display_shopping_list.assert_called_once_with( guid='test_guid', extended=True, ) def test_lists(self): args = shlex.split("lists") show(args) self.mock_display_all_shopping_lists.assert_called_once_with() def test_item_no_guid(self): args = shlex.split("item") self.assertRaises(IndexError, show, args) def test_item_empty_guid(self): args = shlex.split("item ''") self.assertRaises(IndexError, show, args) def test_item(self): args = shlex.split("item test_guid") show(args) self.mock_display_item.assert_called_once_with('test_guid') class TestShowDefaultList: @pytest.fixture(autouse=True) def setUp(self, mocker): self.DEFAULT_LIST_patcher = mock.patch('vt.vt.DEFAULT_LIST', 'default_list') self.DEFAULT_LIST_patcher.start() self.parse_options_patcher = mock.patch('vt.vt.parse_options') self.mock_parse_options = self.parse_options_patcher.start() self.display_shopping_list_patcher = mock.patch('vt.vt.display_shopping_list') self.mock_display_shopping_list = self.display_shopping_list_patcher.start() self.display_all_shopping_lists_patcher = mock.patch( 'vt.vt.display_all_shopping_lists' ) self.mock_display_all_shopping_lists = ( self.display_all_shopping_lists_patcher.start() ) self.display_shopping_list_categories_patcher = mock.patch( 'vt.vt.display_shopping_list_categories' ) self.mock_display_shopping_list_categories = ( self.display_shopping_list_categories_patcher.start() ) mocker.patch.object(term, 'red', autospec=True) self.display_item_patcher = mock.patch('vt.vt.display_item') self.mock_display_item = self.display_item_patcher.start() self.mock_parse_options.return_value = {} yield self.DEFAULT_LIST_patcher.stop() self.parse_options_patcher.stop() self.display_shopping_list_patcher.stop() self.display_all_shopping_lists_patcher.stop() self.display_item_patcher.stop() self.display_shopping_list_categories_patcher.stop() def test_list_empty_guid(self): args = shlex.split("list ''") show(args) self.mock_display_shopping_list.assert_called_once_with(guid='default_list') def test_list_no_guid(self): args = shlex.split("list") show(args) self.mock_display_shopping_list.assert_called_once_with(guid='default_list') def test_list_empty_guid_extended(self): self.mock_parse_options.return_value = {'extended': True} args = shlex.split("list '' -e") show(args) self.mock_display_shopping_list.assert_called_once_with( guid='default_list', extended=True ) def test_list_no_guid_extended(self): self.mock_parse_options.return_value = {'extended': True} args = shlex.split("list -e") show(args) self.mock_display_shopping_list.assert_called_once_with( guid='default_list', extended=True ) def test_list_no_extended(self): args = shlex.split("list test_guid") show(args) self.mock_display_shopping_list.assert_called_once_with(guid='test_guid') def test_list_extended(self): self.mock_parse_options.return_value = {'extended': True} args = shlex.split("list test_guid -e") show(args) self.mock_display_shopping_list.assert_called_once_with( guid='test_guid', extended=True, ) def test_lists(self): args = shlex.split("lists") show(args) self.mock_display_all_shopping_lists.assert_called_once_with() def test_item_no_guid(self): args = shlex.split("item") with pytest.raises(IndexError): show(args) def test_item_empty_guid(self): args = shlex.split("item ''") with pytest.raises(IndexError): show(args) def test_item(self): args = shlex.split("item test_guid") show(args) self.mock_display_item.assert_called_once_with('test_guid') def test_display_list_categories(self): self.mock_parse_options.return_value = { 'categories': [{'name': 'type A'}, {'name': 'type B'}] } args = shlex.split("test_guid") categories(args) self.mock_display_shopping_list_categories.assert_called_once_with('test_guid') def test_display_list_categories_raises(self): self.mock_parse_options.return_value = { 'categories': [{'name': 'type A'}, {'name': 'type B'}] } self.mock_display_shopping_list_categories.side_effect = VittlifyError( 'Got an error' ) args = shlex.split("test_guid") categories(args) term.red.assert_called_once_with('Got an error') self.mock_display_shopping_list_categories.assert_called_once_with('test_guid') def test_display_shopping_list_raises(self): self.mock_display_shopping_list.side_effect = VittlifyError('Got an error') args = shlex.split("list test_guid") show(args) term.red.assert_called_once_with('Got an error') self.mock_display_shopping_list.assert_called_once_with(guid='test_guid') def test_display_item_raises(self): self.mock_display_item.side_effect = VittlifyError('Got an error') args = shlex.split("show test_guid") show(args) term.red.assert_called_once_with('Got an error') def test_display_all_shopping_lists_raises(self): self.mock_display_all_shopping_lists.side_effect = VittlifyError('Got an error') args = shlex.split("lists") show(args) self.mock_display_all_shopping_lists.assert_called_once_with() term.red.assert_called_once_with('Got an error') class TestComplete: @pytest.fixture(autouse=True) def setUp(self, mocker): self.complete_item_patcher = mock.patch('vt.vt.complete_item') self.mock_complete_item = self.complete_item_patcher.start() self.mock_print = mocker.patch('builtins.print') self.display_shopping_list_patcher = mock.patch('vt.vt.display_shopping_list') self.mock_display_shopping_list = self.display_shopping_list_patcher.start() self.apply_strikethrough_patcher = mock.patch('vt.vt.apply_strikethrough') self.mock_apply_strikethrough = self.apply_strikethrough_patcher.start() self.mock_complete_item.return_value = {'name': 'test_name'} self.mock_apply_strikethrough.return_value = 'struck_through' yield self.complete_item_patcher.stop() self.apply_strikethrough_patcher.stop() def test_complete(self): args = shlex.split("test_guid") complete(args) self.mock_complete_item.assert_called_once_with('test_guid', uncomplete=False) self.mock_apply_strikethrough.assert_called_once_with('test_name') self.mock_print.assert_called_once_with( f'Marked {term.magenta}struck_through{term.normal} as done.' ) def test_uncomplete(self): args = shlex.split("test_guid") complete(args, uncomplete=True) self.mock_complete_item.assert_called_once_with('test_guid', uncomplete=True) self.mock_print.assert_called_once_with( f'Marked {term.magenta}test_name{term.normal} undone.' ) def test_done_extended(self): args = shlex.split("-e") complete(args) self.mock_display_shopping_list.assert_called_once_with( extended=True, mode=Status.COMPLETED ) def test_completed_no_extended(self): args = shlex.split("") complete(args) self.mock_display_shopping_list.assert_called_once_with(mode=Status.COMPLETED) def test_completed_extended(self): args = shlex.split("--extended") complete(args) self.mock_display_shopping_list.assert_called_once_with( extended=True, mode=Status.COMPLETED ) class TestModify(unittest.TestCase): def setUp(self): self.modify_item_patcher = mock.patch('vt.vt.modify_item') self.mock_modify_item = self.modify_item_patcher.start() self.display_item_patcher = mock.patch('vt.vt.display_item') self.mock_display_item = self.display_item_patcher.start() def tearDown(self): self.modify_item_patcher.stop() self.display_item_patcher.stop() def test_no_options(self): args = shlex.split("test_guid this is a comment") modify(args) self.mock_modify_item.assert_called_once_with('test_guid', 'this is a comment') self.mock_display_item.assert_called_once_with('test_guid') def test_with_short_options(self): args = shlex.split("test_guid -a this is a comment") modify(args) self.mock_modify_item.assert_called_once_with( 'test_guid', 'this is a comment', append=True ) self.mock_display_item.assert_called_once_with('test_guid') def test_with_options(self): args = shlex.split("test_guid --append this is a comment") modify(args) self.mock_modify_item.assert_called_once_with( 'test_guid', 'this is a comment', append=True ) self.mock_display_item.assert_called_once_with('test_guid') class TestAddDefaultList(unittest.TestCase): def setUp(self): self.DEFAULT_LIST_patcher = mock.patch('vt.vt.DEFAULT_LIST', 'default_list') self.DEFAULT_LIST_patcher.start() self.add_item_patcher = mock.patch('vt.vt.add_item') self.mock_add_item = self.add_item_patcher.start() self.format_row_patcher = mock.patch('vt.vt.format_row') self.mock_format_row = self.format_row_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() def tearDown(self): self.add_item_patcher.stop() self.DEFAULT_LIST_patcher.stop() self.format_row_patcher.stop() self.print_table_patcher.stop() def test_no_guid(self): args = shlex.split("'this is a new item'") add(args) self.mock_add_item.assert_called_once_with('default_list', 'this is a new item') self.mock_format_row.assert_called_once_with( self.mock_add_item.return_value, no_wrap=False ) self.mock_print_table.assert_called_once_with( [self.mock_format_row.return_value] ) def test_with_guid(self): args = shlex.split("test_guid 'this is a new item'") add(args) self.mock_add_item.assert_called_once_with('test_guid', 'this is a new item') self.mock_format_row.assert_called_once_with( self.mock_add_item.return_value, no_wrap=False ) self.mock_print_table.assert_called_once_with( [self.mock_format_row.return_value] ) class TestAddNoDefaultList(unittest.TestCase): def setUp(self): self.DEFAULT_LIST_patcher = mock.patch('vt.vt.DEFAULT_LIST', None) self.DEFAULT_LIST_patcher.start() self.add_item_patcher = mock.patch('vt.vt.add_item') self.mock_add_item = self.add_item_patcher.start() self.format_row_patcher = mock.patch('vt.vt.format_row') self.mock_format_row = self.format_row_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() def tearDown(self): self.add_item_patcher.stop() self.DEFAULT_LIST_patcher.stop() self.format_row_patcher.stop() self.print_table_patcher.stop() def test_no_guid(self): args = shlex.split("'this is a new item'") self.assertRaises(IndexError, add, args) def test_with_guid(self): args = shlex.split("test_guid 'this is a new item'") add(args) self.mock_add_item.assert_called_once_with('test_guid', 'this is a new item') self.mock_format_row.assert_called_once_with( self.mock_add_item.return_value, no_wrap=False ) self.mock_print_table.assert_called_once_with( [self.mock_format_row.return_value] ) class TestMove: @pytest.fixture(autouse=True) def setUp(self, mocker): self.move_item_patcher = mock.patch('vt.vt.move_item') self.mock_move_item = self.move_item_patcher.start() self.mock_print = mocker.patch('builtins.print') yield self.move_item_patcher.stop() def test_(self): args = shlex.split('test_guid to_list_guid') move(args) self.mock_move_item.assert_called_once_with('test_guid', 'to_list_guid') self.mock_print.assert_called_once_with( f'Moved item {term.blue}test_guid{term.normal} to list {term.blue}to_list_guid{term.normal}' ) class TestRun: @pytest.fixture(autouse=True) def setUp(self, mocker): self.show_patcher = mock.patch('vt.vt.show') self.mock_show = self.show_patcher.start() self.complete_patcher = mock.patch('vt.vt.complete') self.mock_complete = self.complete_patcher.start() self.modify_patcher = mock.patch('vt.vt.modify') self.mock_modify = self.modify_patcher.start() self.add_patcher = mock.patch('vt.vt.add') self.mock_add = self.add_patcher.start() self.move_patcher = mock.patch('vt.vt.move') self.mock_move = self.move_patcher.start() mocker.patch.object(term, 'red', autospec=True) self.SHOW_TRACEBACK_patcher = mock.patch('vt.vt.SHOW_TRACEBACK', False) self.SHOW_TRACEBACK_patcher.start() self.PROXY_patcher = mock.patch('vt.vt.PROXY', False) self.PROXY_patcher.start() self.VITTLIFY_URL_patcher = mock.patch('vt.vt.VITTLIFY_URL', 'vittlify_url') self.VITTLIFY_URL_patcher.start() self.help_patcher = mock.patch('vt.vt.help') self.mock_help = self.help_patcher.start() yield self.show_patcher.stop() self.complete_patcher.stop() self.modify_patcher.stop() self.add_patcher.stop() self.move_patcher.stop() self.SHOW_TRACEBACK_patcher.stop() self.PROXY_patcher.stop() self.VITTLIFY_URL_patcher.stop() self.help_patcher.stop() def test_list(self): test_args = shlex.split('list test_guid') run(test_args) self.mock_show.assert_called_once_with(test_args) assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_lists(self): test_args = shlex.split('lists') run(test_args) self.mock_show.assert_called_once_with(test_args) assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_item(self): test_args = shlex.split('item test_guid') run(test_args) self.mock_show.assert_called_once_with(test_args) assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_show(self): test_args = shlex.split('show test_guid') run(test_args) self.mock_show.assert_called_once_with(test_args) assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_done(self): test_args = shlex.split('done test_guid') expected = ['test_guid'] run(test_args) assert not self.mock_show.called self.mock_complete.assert_called_once_with(expected) assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_complete(self): test_args = shlex.split('complete test_guid') expected = ['test_guid'] run(test_args) assert not self.mock_show.called self.mock_complete.assert_called_once_with(expected) assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_undone(self): test_args = shlex.split('undone test_guid') expected = ['test_guid'] run(test_args) assert not self.mock_show.called self.mock_complete.assert_called_once_with(expected, uncomplete=True) assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_uncomplete(self): test_args = shlex.split('uncomplete test_guid') expected = ['test_guid'] run(test_args) assert not self.mock_show.called self.mock_complete.assert_called_once_with(expected, uncomplete=True) assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_modify(self): test_args = shlex.split("modify test_guid 'these are comments'") expected = ['test_guid', 'these are comments'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called self.mock_modify.assert_called_once_with(expected) assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_edit(self): test_args = shlex.split("edit test_guid 'these are comments'") expected = ['test_guid', 'these are comments'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called self.mock_modify.assert_called_once_with(expected) assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_comment(self): test_args = shlex.split("comment test_guid 'these are comments'") expected = ['test_guid', 'these are comments'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called self.mock_modify.assert_called_once_with(expected) assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_comments(self): test_args = shlex.split("comments test_guid 'these are comments'") expected = ['test_guid', 'these are comments'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called self.mock_modify.assert_called_once_with(expected) assert not self.mock_add.called assert not self.mock_move.called assert not self.mock_help.called def test_add(self): test_args = shlex.split("add 'this is a new item'") expected = ['this is a new item'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called assert not self.mock_modify.called self.mock_add.assert_called_once_with(expected) assert not self.mock_move.called assert not self.mock_help.called def test_move(self): test_args = shlex.split("move old_guid new_guid") expected = ['old_guid', 'new_guid'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_help.called self.mock_move.assert_called_once_with(expected) def test_mv(self): test_args = shlex.split("mv old_guid new_guid") expected = ['old_guid', 'new_guid'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_help.called self.mock_move.assert_called_once_with(expected) def test_index_error(self): self.mock_add.side_effect = IndexError() test_args = shlex.split("add 'this is a new item'") with pytest.raises(SystemExit): run(test_args) term.red.assert_called_once_with('Incorrect number of arguments provided') def test_connection_error(self): self.mock_add.side_effect = requests.exceptions.ConnectionError() test_args = shlex.split("add 'this is a new item'") with pytest.raises(SystemExit): run(test_args) term.red.assert_called_once_with( 'Unable to connect to Vittlify instance at vittlify_url' ) def test_http_error(self): self.mock_add.side_effect = requests.exceptions.HTTPError('500 Message') test_args = shlex.split("add 'this is a new item'") with pytest.raises(SystemExit): run(test_args) term.red.assert_called_once_with('Server responded with 500 Message') def test_help(self): test_args = shlex.split("help command") expected = ['command'] run(test_args) assert not self.mock_show.called assert not self.mock_complete.called assert not self.mock_modify.called assert not self.mock_add.called assert not self.mock_move.called self.mock_help.assert_called_once_with(expected) class TestDisplayShoppingListCategories: @pytest.fixture(autouse=True) def setUp(self, mocker): self.get_shopping_list_info_patcher = mock.patch('vt.vt.get_shopping_list_info') self.mock_get_shopping_list_info = self.get_shopping_list_info_patcher.start() self.print_table_patcher = mock.patch('vt.vt.print_table') self.mock_print_table = self.print_table_patcher.start() mocker.patch.object(term, 'red', autospec=True) self.mock_get_shopping_list_info.return_value = {'name': 'test_list'} yield self.get_shopping_list_info_patcher.stop() self.print_table_patcher.stop() def test_no_categories(self): display_shopping_list_categories('test_guid') self.mock_get_shopping_list_info.assert_called_once_with('test_guid') term.red.assert_called_once_with("No categories found for test_list.") def test_has_categories(self): self.mock_get_shopping_list_info.return_value = { 'name': 'test_list', 'categories': [ {'name': 'type A'}, {'name': 'type B'}, ], } display_shopping_list_categories('test_guid') self.mock_print_table.assert_called_once_with( [['type A'], ['type B']], title='test_list' ) class TestHelp(unittest.TestCase): def setUp(self): self.general_help_patcher = mock.patch('vt.vt.GENERAL_HELP') self.mock_general_help = self.general_help_patcher.start() self.lists_help_patcher = mock.patch('vt.vt.LISTS_HELP') self.mock_lists_help = self.lists_help_patcher.start() self.list_help_patcher = mock.patch('vt.vt.LIST_HELP') self.mock_list_help = self.list_help_patcher.start() self.done_help_patcher = mock.patch('vt.vt.DONE_HELP') self.mock_done_help = self.done_help_patcher.start() self.undone_help_patcher = mock.patch('vt.vt.UNDONE_HELP') self.mock_undone_help = self.undone_help_patcher.start() self.comment_help_patcher = mock.patch('vt.vt.COMMENT_HELP') self.mock_comment_help = self.comment_help_patcher.start() self.move_help_patcher = mock.patch('vt.vt.MOVE_HELP') self.mock_move_help = self.move_help_patcher.start() self.categories_help_patcher = mock.patch('vt.vt.CATEGORIES_HELP') self.mock_categories_help = self.categories_help_patcher.start() self.categorize_help_patcher = mock.patch('vt.vt.CATEGORIZE_HELP') self.mock_categorize_help = self.categorize_help_patcher.start() def tearDown(self): self.general_help_patcher.stop() self.lists_help_patcher.stop() self.list_help_patcher.stop() self.done_help_patcher.stop() self.undone_help_patcher.stop() self.comment_help_patcher.stop() self.move_help_patcher.stop() self.categories_help_patcher.stop() self.categorize_help_patcher.stop() def test_no_args(self): expected = self.mock_general_help actual = help([]) self.assertEqual(expected, actual) def test_unknown_command(self): expected = self.mock_general_help actual = help(['unknown command']) self.assertEqual(expected, actual) def test_lists(self): expected = self.mock_lists_help actual = help(['lists']) self.assertEqual(expected, actual) def test_list(self): expected = self.mock_list_help actual = help(['list']) self.assertEqual(expected, actual) def test_done(self): expected = self.mock_done_help actual = help(['done']) self.assertEqual(expected, actual) def test_complete(self): expected = self.mock_done_help actual = help(['complete']) self.assertEqual(expected, actual) def test_undone(self): expected = self.mock_undone_help actual = help(['undone']) self.assertEqual(expected, actual) def test_uncomplete(self): expected = self.mock_undone_help actual = help(['uncomplete']) self.assertEqual(expected, actual) def test_comment(self): expected = self.mock_comment_help actual = help(['comment']) self.assertEqual(expected, actual) def test_modify(self): expected = self.mock_comment_help actual = help(['modify']) self.assertEqual(expected, actual) def test_comments(self): expected = self.mock_comment_help actual = help(['comments']) self.assertEqual(expected, actual) def test_edit(self): expected = self.mock_comment_help actual = help(['edit']) self.assertEqual(expected, actual) def test_move(self): expected = self.mock_move_help actual = help(['move']) self.assertEqual(expected, actual) def test_mv(self): expected = self.mock_move_help actual = help(['mv']) self.assertEqual(expected, actual) def test_categories(self): expected = self.mock_categories_help actual = help(['categories']) self.assertEqual(expected, actual) def test_categorize(self): expected = self.mock_categorize_help actual = help(['categorize']) self.assertEqual(expected, actual) def test_label(self): expected = self.mock_categorize_help actual = help(['label']) self.assertEqual(expected, actual)
[ "kyokley2@gmail.com" ]
kyokley2@gmail.com
6470a8daf591db0dcf2949014f4adbe1d19ddaf5
5d2b04b5454cd3b2d75bf153c96d5fc97e40b2f2
/apps/registro/forms.py
3e23888c85c0dcf3edf5a5897bd25393c443e8ae
[ "Unlicense" ]
permissive
hector-delgado/django-crud-in-docker-container
66162c460c72c6b4bc4d17a13a0e67603f1a5627
6fc78f6b8e88cc9836946cb5f77c3fe39575934e
refs/heads/master
2022-06-18T06:17:25.910698
2020-05-04T17:15:00
2020-05-04T17:15:00
null
0
0
null
null
null
null
UTF-8
Python
false
false
798
py
from django import forms from apps.registro.models import Usuario class UsuarioForm(forms.ModelForm): class Meta: model = Usuario fields = [ 'nombre', 'apellidos', 'edad', 'telefono', 'email', 'domicilio', ] labels = { 'nombre': 'Nombre', 'apellidos': 'Apellidos', 'edad': 'Edad', 'telefono': 'Telefono', 'email': 'E-mail', 'domicilio': 'Domicilio', } widgets = { 'nombre': forms.TextInput(attrs={'class':'form-control'}), 'apellidos': forms.TextInput(attrs={'class':'form-control'}), 'edad': forms.TextInput(attrs={'class':'form-control'}), 'telefono': forms.TextInput(attrs={'class':'form-control'}), 'email': forms.TextInput(attrs={'class':'form-control'}), 'domicilio': forms.TextInput(attrs={'class':'form-control'}), }
[ "hector.delgado.irt@gmail.com" ]
hector.delgado.irt@gmail.com
f34d29f947502d325641f91a58d256f2f869ef4f
c106ec4af915b485d625bfa835f769c2af1d34e3
/trainxgb.py
0f43497d436d2b81bd39167a3f8480667b78e5cf
[]
no_license
Men0x/aobd_project
6820f38f13f2d9ce4c175a035aa4ef30e7925495
d5dcf419775527cfe9025f8d87317ed5a10cd8db
refs/heads/master
2023-01-04T05:34:43.768099
2020-10-30T21:45:06
2020-10-30T21:45:06
306,396,315
0
0
null
null
null
null
UTF-8
Python
false
false
2,296
py
import os import warnings import sys import pandas as pd import numpy as np from sklearn.preprocessing import LabelEncoder from sklearn.metrics import precision_recall_fscore_support as score from sklearn.metrics import accuracy_score from sklearn.model_selection import train_test_split import xgboost as xgb from xgboost import XGBClassifier import mlflow import mlflow.sklearn from preprocessing_data import preprocessing_train, preprocessing_test import logging logging.basicConfig(level=logging.WARN) logger = logging.getLogger(__name__) if __name__ == "__main__": warnings.filterwarnings("ignore") np.random.seed(40) train = preprocessing_train() X = train.drop(columns=['TARGET']) y = train['TARGET'] X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.33, random_state=42) lr = float(sys.argv[1]) if len(sys.argv) > 1 else 0.1 ne = int(sys.argv[2]) if len(sys.argv) > 500 else 100 nj = int(sys.argv[3]) if len(sys.argv) > 3 else 1 with mlflow.start_run(): model = XGBClassifier(learning_rate=lr, n_estimators=ne, n_jobs=nj) model.fit(X_train, y_train) predicted_qualities = model.predict(X_test) precision,recall,fscore,support=score(y_test, predicted_qualities) precision0 = precision[0] precision1 = precision[1] recall0 = recall[0] recall1 = recall[1] fscore0 = fscore[0] fscore1 = fscore[1] support0 = support[0] support1 = support[1] accuracy = accuracy_score(y_test, predicted_qualities) print("XGB Model (learning_rate=%f, n_estimators=%f, n_jobs=%f):" % (lr, ne, nj)) mlflow.log_param("learning_rate", lr) mlflow.log_param("n_estimators", ne) mlflow.log_param("n_jobs", nj) mlflow.log_metric("precision0", precision0) mlflow.log_metric("precision1", precision1) mlflow.log_metric("recall0", recall0) mlflow.log_metric("recall1", recall1) mlflow.log_metric("fscore0", fscore0) mlflow.log_metric("fscore1", fscore1) mlflow.log_metric("support0", support0) mlflow.log_metric("support1", support1) mlflow.log_metric("accuracy", accuracy) mlflow.sklearn.log_model(model, "XGBoostClassifier")
[ "38214356+Men0x@users.noreply.github.com" ]
38214356+Men0x@users.noreply.github.com
54d614a2458bb581353bf4aeba881b56472296c9
4d69d32a3dd1c45d3e0dafc9bacd4f1de5c6ea6d
/tests/handlers/test_sync.py
a01ab471f5944031638f4f1ffee6f8dc07131890
[ "Apache-2.0" ]
permissive
codemonk-sunhui/synapse
6fd1ab9359cbd0c84852cbce5c65858d98c60e66
7824a751a769e692113d0dc19fe9a3ba99007940
refs/heads/master
2020-03-27T06:52:49.273620
2018-11-23T10:19:46
2018-11-23T10:19:46
146,142,219
0
0
Apache-2.0
2018-08-26T01:33:32
2018-08-26T01:33:32
null
UTF-8
Python
false
false
2,628
py
# -*- coding: utf-8 -*- # Copyright 2018 New Vector Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from twisted.internet import defer from synapse.api.errors import Codes, ResourceLimitError from synapse.api.filtering import DEFAULT_FILTER_COLLECTION from synapse.handlers.sync import SyncConfig, SyncHandler from synapse.types import UserID import tests.unittest import tests.utils from tests.utils import setup_test_homeserver class SyncTestCase(tests.unittest.TestCase): """ Tests Sync Handler. """ @defer.inlineCallbacks def setUp(self): self.hs = yield setup_test_homeserver(self.addCleanup) self.sync_handler = SyncHandler(self.hs) self.store = self.hs.get_datastore() @defer.inlineCallbacks def test_wait_for_sync_for_user_auth_blocking(self): user_id1 = "@user1:server" user_id2 = "@user2:server" sync_config = self._generate_sync_config(user_id1) self.hs.config.limit_usage_by_mau = True self.hs.config.max_mau_value = 1 # Check that the happy case does not throw errors yield self.store.upsert_monthly_active_user(user_id1) yield self.sync_handler.wait_for_sync_for_user(sync_config) # Test that global lock works self.hs.config.hs_disabled = True with self.assertRaises(ResourceLimitError) as e: yield self.sync_handler.wait_for_sync_for_user(sync_config) self.assertEquals(e.exception.errcode, Codes.RESOURCE_LIMIT_EXCEED) self.hs.config.hs_disabled = False sync_config = self._generate_sync_config(user_id2) with self.assertRaises(ResourceLimitError) as e: yield self.sync_handler.wait_for_sync_for_user(sync_config) self.assertEquals(e.exception.errcode, Codes.RESOURCE_LIMIT_EXCEED) def _generate_sync_config(self, user_id): return SyncConfig( user=UserID(user_id.split(":")[0][1:], user_id.split(":")[1]), filter_collection=DEFAULT_FILTER_COLLECTION, is_guest=False, request_key="request_key", device_id="device_id", )
[ "neil@matrix.org" ]
neil@matrix.org
c0f8ac78473e82ea1426614dacf74fc4180ef9e9
dd44f2823b7ea9c8e20dce41844627d9371989d8
/macroPCA-old.py
2495a1fe23fc8434df5096a63eaf432a0cd4f3a4
[]
no_license
jmfreeland/macroAnalysis
affa7a3432f76bcee3d968bfd3ec701a7b3c1871
4a0e551832f4f2de7fbd2e4060e69ae7e7096068
refs/heads/master
2022-12-05T01:28:11.669592
2020-08-27T08:04:39
2020-08-27T08:04:39
257,497,581
0
0
null
null
null
null
UTF-8
Python
false
false
2,371
py
# -*- coding: utf-8 -*- """ Macro PCA: Objective: Identify high-level drivers of movement in a collection of macro-sensitive tradable securities outputs: PCA loadings, time-series of factor movements todo: try with fractionally differenced series because why not -Find stationary series for factor 1 and adjust risk appetite accordingly -Find trends for all factors, invest in alighment with trends . - """ #import required models import pandas as pd import yfinance as yf import numpy as np from sklearn.decomposition import PCA import seaborn as sns import matplotlib.pyplot as plt #set tickers for data gathering, collect data #test_stocks = ['SPY', 'TLT', 'IEF','XLU','IGV','QQQ','GLD','GDX','MBB','XBI','MOO','TIP','EEM','UUP'] test_stocks = ['SPY','QQQ','TLT','IEF','SHY','TIP','AGG','LQD','IWM','IVW','IVE','IYF','IYE','IYM'] tickerData = {} tickerDF = {} for ticker in test_stocks: tickerData[ticker] = yf.Ticker(ticker) #get the historical prices for this ticker tickerDF[ticker] = tickerData[ticker].history(period='1d', start='2000-1-1', end='2020-12-31') #choose column on which to run PCA, organize a historical table for analysis test_col = 'Close' pca_data = pd.DataFrame(columns=test_stocks) for ticker in test_stocks: pca_data.loc[:,ticker] = tickerDF[ticker].loc[:,test_col] #diagnostic - see when your series begin for ticker in test_stocks: print(ticker , " " , tickerDF[ticker].index[1]) #create pca form sklearn and run it on outrights pca = PCA(n_components=len(test_stocks)-1) test_data = pca_data.dropna() outright_pca = pca.fit(test_data) outright_loadings = outright_pca.components_ outright_variances = outright_pca.explained_variance_ outright_stdev = np.sqrt(outright_variances) #for i in range(0,outright_loadings.shape[0]): for i in range(0,5): fig = plt.figure(figsize=(16,9), dpi=300) fig.suptitle(('Macro PCA loadings: factor ' + str(i+1))) sns.barplot(x=test_stocks,y=outright_loadings[i]) outright_time_series = pd.DataFrame(outright_pca.transform(test_data), index=test_data.index) outright_time_series.columns = outright_time_series.columns+1 fig = plt.figure(figsize=(16,9), dpi=300) time_plot = sns.lineplot(data=outright_time_series, dashes=False) fig = plt.figure(figsize=(16,9), dpi=300) time_plot = sns.lineplot(data=outright_time_series.iloc[:,1:], dashes=False)
[ "josh.freeland@gmail.com" ]
josh.freeland@gmail.com
f589f57cc509b15cda51ec3667766cf1f98dff37
988dc7a525c26cc68ce7f64ddcdc77bce0995231
/lists/tests/test_views.py
a5b3c6df929b5df087c1f226335ace45d16ea0b4
[]
no_license
bohlool/django-tdd-book
32be2f7f0ecaa04b6aca040778fc284079179a0e
0f086df4151176585587555533f41a7754b0bb16
refs/heads/master
2023-03-18T01:50:47.545168
2018-08-01T08:40:51
2018-08-01T08:40:51
null
0
0
null
null
null
null
UTF-8
Python
false
false
8,810
py
from django.test import TestCase from django.utils.html import escape from lists.models import Item, List from django.contrib.auth import get_user_model from lists.forms import ( DUPLICATE_ITEM_ERROR, EMPTY_ITEM_ERROR, ExistingListItemForm, ItemForm#, NewListForm ) from unittest.mock import patch, Mock from django.http import HttpRequest import unittest from lists.views import new_list User = get_user_model() # Create your tests here. class HomePageTest(TestCase): def test_uses_home_template(self): response = self.client.get('/') self.assertTemplateUsed(response, 'home.html') def test_home_page_uses_item_form(self): response = self.client.get('/') #print(response.context['form']) self.assertIsInstance(response.context['form'], ItemForm) class NewListViewIntegratedTest(TestCase): def test_can_save_a_POST_request(self): self.client.post('/lists/new', data={'text': 'A new list item'}) self.assertEqual(Item.objects.count(), 1) new_item = Item.objects.first() self.assertEqual(new_item.text, 'A new list item') def test_list_owner_is_saved_if_user_is_authenticated(self): user = User.objects.create(email='a@b.com') self.client.force_login(user) self.client.post('/lists/new', data={'text': 'new item'}) list_ = List.objects.first() self.assertEqual(list_.owner, user) def test_for_invalid_input_doesnt_save_but_shows_errors(self): response = self.client.post('/lists/new', data={'text': ''}) self.assertEqual(List.objects.count(), 0) self.assertContains(response, escape(EMPTY_ITEM_ERROR)) @patch('lists.views.NewListForm') class NewListViewUnitTest(unittest.TestCase): def setUp(self): self.request = HttpRequest() self.request.POST['text'] = 'new list item' self.request.user = Mock() def test_passes_POST_data_to_NewListForm(self, mockNewListForm): new_list(self.request) mockNewListForm.assert_called_once_with(data=self.request.POST) def test_saves_form_with_owner_if_form_valid(self, mockNewListForm): mock_form = mockNewListForm.return_value mock_form.is_valid.return_value = True new_list(self.request) mock_form.save.assert_called_once_with(owner=self.request.user) def test_does_not_save_if_form_invalid(self, mockNewListForm): mock_form = mockNewListForm.return_value mock_form.is_valid.return_value = False new_list(self.request) self.assertFalse(mock_form.save.called) @patch('lists.views.render') def test_renders_home_template_with_form_if_form_invalid( self, mock_render, mockNewListForm ): mock_form = mockNewListForm.return_value mock_form.is_valid.return_value = False response = new_list(self.request) self.assertEqual(response, mock_render.return_value) mock_render.assert_called_once_with( self.request, 'home.html', {'form': mock_form} ) @patch('lists.views.redirect') def test_redirects_to_form_returned_object_if_form_valid( self, mock_redirect, mockNewListForm ): mock_form = mockNewListForm.return_value mock_form.is_valid.return_value = True response = new_list(self.request) self.assertEqual(response, mock_redirect.return_value) mock_redirect.assert_called_once_with(mock_form.save.return_value) class ListViewTest(TestCase): def test_uses_list_template(self): list_ = List.objects.create() response = self.client.get(f'/lists/{list_.id}/') self.assertTemplateUsed(response, 'list.html') def test_passes_correct_list_to_template(self): other_list = List.objects.create() correct_list = List.objects.create() response = self.client.get(f'/lists/{correct_list.id}/') self.assertEqual(response.context['list'], correct_list) del(other_list) # getting warnings about unused variables, fixing def test_displays_only_items_for_that_list(self): correct_list = List.objects.create() Item.objects.create(text='itemey 1', list=correct_list) Item.objects.create(text='itemey 2', list=correct_list) other_list = List.objects.create() Item.objects.create(text='other list item 1', list=other_list) Item.objects.create(text='other list item 2', list=other_list) response = self.client.get(f'/lists/{correct_list.id}/') self.assertContains(response, 'itemey 1') self.assertContains(response, 'itemey 2') self.assertNotContains(response, 'other list item 1') self.assertNotContains(response, 'other list item 2') def test_can_save_a_POST_request_to_an_existing_list(self): other_list = List.objects.create() correct_list = List.objects.create() self.client.post( f'/lists/{correct_list.id}/', data={'text': 'A new item for an existing list'} ) self.assertEqual(Item.objects.count(), 1) new_item = Item.objects.first() self.assertEqual(new_item.text, 'A new item for an existing list') self.assertEqual(new_item.list, correct_list) del(other_list) # getting warnings about unused variables, fixing def test_POST_redirects_to_list_view(self): other_list = List.objects.create() correct_list = List.objects.create() response = self.client.post( f'/lists/{correct_list.id}/', data={'text' : 'A new item for an existing list'} ) self.assertRedirects(response, f'/lists/{correct_list.id}/') del(other_list) # getting warnings about unused variables, fixing def test_displays_item_form(self): list_ = List.objects.create() response = self.client.get(f'/lists/{list_.id}/') self.assertIsInstance(response.context['form'], ExistingListItemForm) self.assertContains(response, 'name="text"') def post_invalid_input(self): list_ = List.objects.create() return self.client.post( f'/lists/{list_.id}/', data={'text': ''} ) def test_for_invalid_input_nothing_saved_to_db(self): self.post_invalid_input() self.assertEqual(Item.objects.count(), 0) def test_for_invalid_input_renders_list_template(self): response = self.post_invalid_input() self.assertEqual(response.status_code, 200) self.assertTemplateUsed(response, 'list.html') def test_for_invalid_input_passes_form_to_template(self): response = self.post_invalid_input() self.assertIsInstance(response.context['form'], ItemForm) def test_for_invalid_input_shows_error_on_page(self): response = self.post_invalid_input() self.assertContains(response, escape(EMPTY_ITEM_ERROR)) def test_duplicate_item_validation_errors_end_up_on_lists_page(self): list1 = List.objects.create() item1 = Item.objects.create(list=list1, text='textey') response = self.client.post( f'/lists/{list1.id}/', data={'text': 'textey'} ) expected_error = escape(DUPLICATE_ITEM_ERROR) self.assertContains(response, expected_error) self.assertTemplateUsed(response, 'list.html') self.assertEqual(Item.objects.all().count(), 1) del(item1) class MyListsTest(TestCase): def test_my_lists_url_renders_my_lists_template(self): User.objects.create(email='a@b.com') response = self.client.get('/lists/users/a@b.com/') self.assertTemplateUsed(response, 'my_lists.html') def test_passes_correct_owner_to_template(self): User.objects.create(email='wrong@owner.com') correct_user = User.objects.create(email='a@b.com') response = self.client.get('/lists/users/a@b.com/') self.assertEqual(response.context['owner'], correct_user) class SharingTest(TestCase): def test_post_redirects_to_list_page(self): list_ = List.objects.create() response = self.client.post( f'/lists/{list_.id}/share', data={'sharee': 'share@ab.com'} ) #self.assertEqual(response.status_code, 302) self.assertRedirects(response, list_.get_absolute_url()) def test_created_user_appears_on_shared_list(self): sharee = User.objects.create(email='sharee@ab.com') # do we need to get a logged in user? list_ = List.objects.create() self.client.post( f'/lists/{list_.id}/share', data={'sharee': sharee.email} ) self.assertIn(sharee, list_.shared_with.all())
[ "ben_jacobson@live.com" ]
ben_jacobson@live.com
a6630d2aeb6791c9a89d193f911eb7ef74dbe858
b9da0a46452a67d4d2e1c6e76d982295b8e9bec4
/double_joint_arm/robot/robot_state_ideal_speed.py
402873c81f5f1c2078042928112a15291143bf68
[]
no_license
lessthantrue/RobotProjects
d68d0cfba090ca0e117c78baae7afdc13203ec6c
e84069ad90dca9e3d4873febe1bcb40f396d656e
refs/heads/master
2023-02-06T00:02:02.070109
2020-12-27T06:00:33
2020-12-27T06:00:33
291,508,533
3
0
null
null
null
null
UTF-8
Python
false
false
1,331
py
import numpy as np import kinematics from matrix_utils import * def close(v1, v2, tol=0.0001): return abs(np.linalg.norm(v1-v2)) < tol class Robot(): def __init__(self, t1_0, t2_0, x0=1, y0=1): self.len1 = 1 self.len2 = 2 # frames self.Rw_0 = translation2d(x0, y0) self.R0_1_base = translation2d(self.len1, 0) self.R1_2_base = translation2d(self.len2, 0) self.R0_1 = rotation2d(t1_0) @ self.R0_1_base self.R1_2 = rotation2d(t2_0) @ self.R1_2_base self.x0, self.y0 = x0, y0 self.t1, self.t2 = t1_0, t2_0 self.w1, self.w2 = 0, 0 # c1, c2 = w1, w2 def act(self, c1, c2, dt): self.t1 += (self.w1 + c1) / 2 * dt self.t2 += (self.w2 + c2) / 2 * dt self.w1 = c1 self.w2 = c2 self.R0_1 = rotation2d(self.t1) @ self.R0_1_base self.R1_2 = rotation2d(self.t2) @ self.R1_2_base def getStateVector(self): return np.array([self.t1, self.t2, self.w1, self.w2]) def getVelVector(self): return np.array([self.w1, self.w2, 0, 0]) def getJoints(self): q0 = self.Rw_0 @ np.array([0, 0, 1]) q1 = self.Rw_0 @ self.R0_1 @ np.array([0, 0, 1]) q2 = self.Rw_0 @ self.R0_1 @ self.R1_2 @ np.array([0, 0, 1]) return (q0, q1, q2)
[ "nickm@outofthisworld.net" ]
nickm@outofthisworld.net
b09e483727dc8717cd45a0f85c80edd42152255e
ec09ccf35bcc968b7b1b6cd2aec5a150ecbc2e01
/test/hpa_test.py
475e98ef79603c23cd3ba21807622bbec5f1fd3a
[ "CC0-1.0" ]
permissive
ssupdoc/k8-simulation
1717b35d8aa1b6ab0076ceadacdfe475f81afcd3
7834d3faaed3e86b547554c6228540c316621011
refs/heads/master
2022-12-23T21:58:27.083986
2020-10-04T01:05:24
2020-10-04T01:05:24
285,696,159
1
0
CC0-1.0
2020-10-04T01:05:25
2020-08-07T00:03:20
Python
UTF-8
Python
false
false
1,068
py
from src.api_server import APIServer from src.load_balancer import LoadBalancer from src.hpa import HPA from src.pod import Pod import unittest DEPLOYMENT_INFO = ['Deployment_AA', 2, 2] HPA_INFO = ['Deployment_AA', 75, 10, 5] _hpaCtlLoop = 2 apiServer = APIServer() apiServer.CreateDeployment(DEPLOYMENT_INFO) deployment = apiServer.etcd.deploymentList[0] podName = deployment.deploymentLabel + "_" + str(apiServer.GeneratePodName()) pod = Pod(podName, deployment.cpuCost, deployment.deploymentLabel) pod.status = "RUNNING" pod.requests = [ 'Req 1' ] pod.available_cpu -= 1 podList = [pod, pod] hpa = HPA(apiServer, _hpaCtlLoop, HPA_INFO) class TestUtilisation(unittest.TestCase): def test_average_utilisation(self): load = hpa.calculateAvgUtil(deployment, podList) self.assertEqual(load, 0.5) class TestController(unittest.TestCase): def test_controller_update(self): hpa.updateController(10, 12) self.assertEqual(hpa.controller.kp, 10) self.assertEqual(hpa.controller.ki, 12) self.assertEqual(hpa.pValue, 10) self.assertEqual(hpa.iValue, 12)
[ "sriramemailsyou@gmail.com" ]
sriramemailsyou@gmail.com
5da1e0d740b5a0dba162b287dd9c836577381ca3
753410fcbb13267827464af08120543279a4d7dc
/ride_height.py
5c20d103670047c48b5352f7cd5e4c0247ce6284
[]
no_license
donour/sharks
93b18c34419bbd408d5022188f3f7baae233785c
d243408cb68cda18f8121d39ada393d49a01af91
refs/heads/master
2020-06-02T07:50:35.925352
2013-11-16T21:07:02
2013-11-16T21:07:02
null
0
0
null
null
null
null
UTF-8
Python
false
false
1,157
py
#!/usr/bin/python import quick2wire.i2c as i2c bus = i2c.I2CMaster() adc_address1 = 0x68 adc_address2 = 0x69 varDivisior = 16 # from pdf sheet on adc addresses and config varMultiplier = (2.4705882/varDivisior)/1000 def changechannel(address, adcConfig): bus.transaction(i2c.writing_bytes(address, adcConfig)) def getadcreading(address): h, m, l ,s = bus.transaction(i2c.reading(address,4))[0] t = h << 8 | m # check if positive or negative number and invert if needed if (h > 128): t = ~(0x020000 - t) return t * varMultiplier def setadc(addr): mode = 1 sr = 2 # 0:240, 1:60, 2:15, 3:3.75 gain = 0 # gain = 2^x config_register = 0; config_register |= 0 << 5 config_register |= mode << 4 config_register |= sr << 2 config_register |= gain bus.transaction(i2c.writing_bytes(addr, config_register)) start = 0.0 setadc(adc_address1) changechannel(adc_address2, 0x9C) def height(): return getadcreading(adc_address1) if __name__ == "__main__": import sys,time while True: s = "\r%.6f" % height() sys.stdout.write(s) time.sleep(0.1)
[ "donour@cs.unm.edu" ]
donour@cs.unm.edu
ea5f7a09a051a8dfce6a47bf73c4c80e1df83072
146e54c9b4654ae24cf61e7bdc63ec974034b6a4
/kafka/getFromSQS_kafka.py
10907cec0257a337a69700a4b20cef6cbca0af78
[]
no_license
sajal50/twittTrends
094e86712428e4fc394bb698a70c5574485bb09c
e2b1c2e1ab2f71c14a78a2bc2e37e6282898133a
refs/heads/master
2020-06-20T12:08:47.926582
2017-01-15T12:26:02
2017-01-15T12:26:02
74,866,195
0
1
null
null
null
null
UTF-8
Python
false
false
1,599
py
from multiprocessing import Pool, TimeoutError, Lock import threading import time import os import multiprocessing from watson_developer_cloud import AlchemyLanguageV1 import boto3 import config as Config #Using config file to read the config settings from a separate config file. from kafka import KafkaConsumer import json KAFKA_HOST = 'localhost:9092' TOPIC = 'test' lock = Lock() alchemy_language = AlchemyLanguageV1(api_key=Config.ALCHEMY_API_KEY) def getFromKafka(): lock.acquire() consumer = KafkaConsumer(TOPIC, bootstrap_servers=[KAFKA_HOST],consumer_timeout_ms=10000) for message in consumer: tweet = json.loads(message.value) if tweet is not None : #getting sentiment result = json.loads(json.dumps( alchemy_language.sentiment( text = tweet['text']), indent = 2)) docSentiment = result['docSentiment'] sentiment = docSentiment['type'] tweet['sentiment'] = sentiment tweet = json.dumps(tweet) print tweet client = boto3.client('sns') response = client.publish( TargetArn=Config.SNS_ARN, Message=json.dumps({'default': tweet}), MessageStructure='json' ) lock.release() if __name__ == '__main__': #getFromKafka() pool = Pool(processes=4) # start 4 worker processes while 1: multiple_results = [pool.apply_async(getFromKafka, ()) for i in range(4)] print [res.get(timeout=100) for res in multiple_results] time.sleep(2)
[ "sajal50@gmail.com" ]
sajal50@gmail.com