commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
f02f8f5a68bd26d1ece32c50482729b7774b6e2a
scripts/simple-analysis.py
scripts/simple-analysis.py
#!/usr/bin/python from __future__ import print_function, division import networkx from reddit_meatspace.models import MeetupConnections connections = MeetupConnections._byID("2013") digraph = networkx.DiGraph() for connection, timestamp in connections._values().iteritems(): left, right = connection.split(":") digraph.add_edge(left, right) lenient = digraph.to_undirected(reciprocal=False) strict = digraph.to_undirected(reciprocal=True) meetups = networkx.connected_component_subgraphs(lenient) print("{0} people @ {1} meetups (avg. {2:.2} per meetup)".format( len(lenient), len(meetups), len(lenient) / len(meetups))) print("{0} connections of {1} distinct meetings ({2:.2%})".format(strict.size(), lenient.size(), strict.size() / lenient.size()))
Add a simple script for looking at connections.
Add a simple script for looking at connections.
Python
bsd-3-clause
reddit/reddit-plugin-meatspace,reddit/reddit-plugin-meatspace,reddit/reddit-plugin-meatspace
--- +++ @@ -0,0 +1,22 @@ +#!/usr/bin/python + +from __future__ import print_function, division + +import networkx + +from reddit_meatspace.models import MeetupConnections + + +connections = MeetupConnections._byID("2013") +digraph = networkx.DiGraph() +for connection, timestamp in connections._values().iteritems(): + left, right = connection.split(":") + digraph.add_edge(left, right) + +lenient = digraph.to_undirected(reciprocal=False) +strict = digraph.to_undirected(reciprocal=True) +meetups = networkx.connected_component_subgraphs(lenient) + +print("{0} people @ {1} meetups (avg. {2:.2} per meetup)".format( + len(lenient), len(meetups), len(lenient) / len(meetups))) +print("{0} connections of {1} distinct meetings ({2:.2%})".format(strict.size(), lenient.size(), strict.size() / lenient.size()))
39b63523634801fe8ef2cca03e11b3875d84cdbd
flare/flare_io.py
flare/flare_io.py
from flare.struc import Structure from typing import List from json import dump, load from flare.util import NumpyEncoder def md_trajectory_to_file(filename, structures: List[Structure]): """ Take a list of structures and write them to a json file. :param filename: :param structures: """ f = open(filename, 'w') dump([s.as_dict() for s in structures], f, cls=NumpyEncoder) f.close() def md_trajectory_from_file(filename): """ Read a list of structures from a json file, formatted as in md_trajectory_to_file. :param filename: """ f = open(filename, 'r') structure_list = load(f) structures = [Structure.from_dict(dictionary) for dictionary in structure_list] return structures
from flare.struc import Structure from typing import List from json import dump, load from flare.util import NumpyEncoder def md_trajectory_to_file(filename: str, structures: List[Structure]): """ Take a list of structures and write them to a json file. :param filename: :param structures: """ with open(filename, 'w') as f: dump([s.as_dict() for s in structures], f, cls=NumpyEncoder) def md_trajectory_from_file(filename: str): """ Read a list of structures from a json file, formatted as in md_trajectory_to_file. :param filename: """ with open(filename, 'r') as f: structure_list = load(f) structures = [Structure.from_dict(dictionary) for dictionary in structure_list] return structures
Tweak syntax for f.close() concision, add typehints
Tweak syntax for f.close() concision, add typehints
Python
mit
mir-group/flare,mir-group/flare
--- +++ @@ -3,22 +3,21 @@ from json import dump, load from flare.util import NumpyEncoder -def md_trajectory_to_file(filename, structures: List[Structure]): +def md_trajectory_to_file(filename: str, structures: List[Structure]): """ Take a list of structures and write them to a json file. :param filename: :param structures: """ - f = open(filename, 'w') - dump([s.as_dict() for s in structures], f, cls=NumpyEncoder) - f.close() + with open(filename, 'w') as f: + dump([s.as_dict() for s in structures], f, cls=NumpyEncoder) -def md_trajectory_from_file(filename): +def md_trajectory_from_file(filename: str): """ Read a list of structures from a json file, formatted as in md_trajectory_to_file. :param filename: """ - f = open(filename, 'r') - structure_list = load(f) - structures = [Structure.from_dict(dictionary) for dictionary in structure_list] + with open(filename, 'r') as f: + structure_list = load(f) + structures = [Structure.from_dict(dictionary) for dictionary in structure_list] return structures
cb0d6124ea31e8fb9ff8957072a2b881b882127e
examples/hero9_timelapse_webcam.py
examples/hero9_timelapse_webcam.py
import sys import time from goprocam import GoProCamera, constants import threading import logging """ I use PM2 to start my GoPro cameras, using a Raspberry Pi 4, works perfectly. pm2 start timelapse.py --cron "30 7 * * *" --log timelapse.log --no-autorestart This script will overrride some settings for reliability: Voice control: OFF AutoPower off: NEVER Beeps: OFF (Do not want the camera beeping at 6AM) NightLapse configuration left untouched, I recommend always using Auto shutter for sunrise and locking the White Balance to 4000k or higher. """ def start_timelapse(interface): gopro = GoProCamera.GoPro(ip_address=GoProCamera.GoPro.getWebcamIP( interface), camera=constants.gpcontrol, webcam_device=interface) logging.info( "Started goprocam instance with interface {}".format(interface)) gopro.gpControlSet(constants.Setup.VOICE_CONTROL, constants.Setup.VoiceControl.OFF) gopro.gpControlSet(constants.Setup.AUTO_OFF, constants.Setup.AutoOff.Never) logging.info("All config set") gopro.mode(constants.Mode.MultiShotMode, constants.Mode.SubMode.MultiShot.NightLapse) gopro.shutter(constants.start) logging.info("Started timelapse") cameras = sys.argv[1] cameras = cameras.split(",") for interface in cameras: thr = threading.Thread(target=start_timelapse, args=(interface,)) thr.start()
Add Timelapse script for sunrise timelapses
Add Timelapse script for sunrise timelapses
Python
mit
KonradIT/gopro-py-api,KonradIT/gopro-py-api
--- +++ @@ -0,0 +1,42 @@ +import sys +import time +from goprocam import GoProCamera, constants +import threading +import logging + +""" +I use PM2 to start my GoPro cameras, using a Raspberry Pi 4, works perfectly. + +pm2 start timelapse.py --cron "30 7 * * *" --log timelapse.log --no-autorestart + +This script will overrride some settings for reliability: + +Voice control: OFF +AutoPower off: NEVER +Beeps: OFF (Do not want the camera beeping at 6AM) + +NightLapse configuration left untouched, I recommend always using Auto shutter for sunrise and locking the White Balance to 4000k or higher. +""" + + +def start_timelapse(interface): + gopro = GoProCamera.GoPro(ip_address=GoProCamera.GoPro.getWebcamIP( + interface), camera=constants.gpcontrol, webcam_device=interface) + logging.info( + "Started goprocam instance with interface {}".format(interface)) + gopro.gpControlSet(constants.Setup.VOICE_CONTROL, + constants.Setup.VoiceControl.OFF) + gopro.gpControlSet(constants.Setup.AUTO_OFF, constants.Setup.AutoOff.Never) + logging.info("All config set") + gopro.mode(constants.Mode.MultiShotMode, + constants.Mode.SubMode.MultiShot.NightLapse) + gopro.shutter(constants.start) + logging.info("Started timelapse") + + +cameras = sys.argv[1] +cameras = cameras.split(",") + +for interface in cameras: + thr = threading.Thread(target=start_timelapse, args=(interface,)) + thr.start()
4651e178ddbeac9211f8170e2e20f8a35ff0e3ab
ocradmin/plugins/test_nodetree.py
ocradmin/plugins/test_nodetree.py
#!/usr/bin/python import os import sys import json sys.path.append(os.path.abspath("..")) os.environ['DJANGO_SETTINGS_MODULE'] = 'ocradmin.settings' sys.path.insert(0, "lib") from nodetree import script from nodetree.manager import ModuleManager def run(nodelist, outpath): manager = ModuleManager() manager.register_module("ocradmin.plugins.ocropus_nodes") manager.register_module("ocradmin.plugins.tesseract_nodes") manager.register_module("ocradmin.plugins.cuneiform_nodes") manager.register_module("ocradmin.plugins.numpy_nodes") manager.register_module("ocradmin.plugins.pil_nodes") s = script.Script(nodelist, manager=manager) term = s.get_terminals()[0] print "Rendering to %s" % outpath out = manager.get_new_node("Ocropus::FileOut", label="Output", params=[("path", os.path.abspath(outpath))]) out.set_input(0, term) out.eval() if __name__ == "__main__": if len(sys.argv) < 3: print "Usage: %s <script> <output>" % sys.argv[0] sys.exit(1) nodes = None with open(sys.argv[1], "r") as f: nodes = json.load(f) if nodes is None: print "No nodes found in script" sys.exit(1) run(nodes, sys.argv[2])
Add a simple CLI script for writing out results of scripts
Add a simple CLI script for writing out results of scripts
Python
apache-2.0
vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium,vitorio/ocropodium
--- +++ @@ -0,0 +1,44 @@ +#!/usr/bin/python + +import os +import sys +import json +sys.path.append(os.path.abspath("..")) +os.environ['DJANGO_SETTINGS_MODULE'] = 'ocradmin.settings' + +sys.path.insert(0, "lib") + +from nodetree import script +from nodetree.manager import ModuleManager + +def run(nodelist, outpath): + manager = ModuleManager() + manager.register_module("ocradmin.plugins.ocropus_nodes") + manager.register_module("ocradmin.plugins.tesseract_nodes") + manager.register_module("ocradmin.plugins.cuneiform_nodes") + manager.register_module("ocradmin.plugins.numpy_nodes") + manager.register_module("ocradmin.plugins.pil_nodes") + + s = script.Script(nodelist, manager=manager) + term = s.get_terminals()[0] + print "Rendering to %s" % outpath + out = manager.get_new_node("Ocropus::FileOut", label="Output", + params=[("path", os.path.abspath(outpath))]) + out.set_input(0, term) + out.eval() + +if __name__ == "__main__": + if len(sys.argv) < 3: + print "Usage: %s <script> <output>" % sys.argv[0] + sys.exit(1) + + nodes = None + with open(sys.argv[1], "r") as f: + nodes = json.load(f) + + if nodes is None: + print "No nodes found in script" + sys.exit(1) + + run(nodes, sys.argv[2]) +
23138ab91e5ac0ecf92a0968bf8e4abfa7d0c763
removedups.py
removedups.py
import hashlib, csv, os def md5(fname): hash_md5 = hashlib.md5() with open(fname, "rb") as f: for chunk in iter(lambda: f.read(4096), b""): hash_md5.update(chunk) return hash_md5.hexdigest() def process_directory_csv(current_dir_fullpath, sub_dir_list, files, csvwriter): for file in files: full_name = current_dir_fullpath + '/' + file # print(" " + full_name) csvwriter.writerow([md5(full_name), str(os.path.getsize(full_name)), full_name]) def walk_all_subdirectories(path, output_file_name): # count = 0 with open(output_file_name, "w") as csvfile: csvwriter = csv.writer(csvfile, delimiter=':', quotechar='|', quoting=csv.QUOTE_MINIMAL) for current_dir, sub_dirs, files in os.walk(path): print(current_dir) process_directory_csv(current_dir, sub_dirs, files, csvwriter) csvfile.flush() # DEBUG CODE - process only 5 directories # count += 1 # if count >= 10: # csvfile.close() # break; csvfile.close() def sort_file(inname, outname): input_file = open(inname, "r") output_file = open(outname, "w", 1) lines = [] # give lines variable a type of list for line in input_file: lines.append(line) lines.sort() for line in lines: output_file.write(line) input_file.close() output_file.close() def generate_delete_commands(sortedfile, outname): import csv output_file = open(outname, "w", 1) previous_checksum = "IMPOSSIBLE_CHECKSUM" with open(sortedfile) as f: reader = csv.reader(f, delimiter=':', quotechar='|', quoting=csv.QUOTE_MINIMAL) for row in reader: # print(row[0], row) if previous_checksum == row[0]: output_file.write("rm '" + row[2] + "'\n") print("removing " + row[2]) os.remove(row[2]) previous_checksum = row[0] f.close() output_file.close() # Main program follows directory_name = ".." unsorted_file_name = "filelist.csv" sorted_file_name = "sortedfilelist.csv" delete_command_file_name = "deletecommands.sh" if __name__ == '__main__': walk_all_subdirectories('..', unsorted_file_name) sort_file(unsorted_file_name, sorted_file_name) generate_delete_commands(sorted_file_name, delete_command_file_name)
Remove duplicates in all subdirectories - working raw version.
Remove duplicates in all subdirectories - working raw version.
Python
apache-2.0
alprab/utils
--- +++ @@ -0,0 +1,70 @@ +import hashlib, csv, os + +def md5(fname): + hash_md5 = hashlib.md5() + with open(fname, "rb") as f: + for chunk in iter(lambda: f.read(4096), b""): + hash_md5.update(chunk) + return hash_md5.hexdigest() + +def process_directory_csv(current_dir_fullpath, sub_dir_list, files, csvwriter): + for file in files: + full_name = current_dir_fullpath + '/' + file + # print(" " + full_name) + csvwriter.writerow([md5(full_name), str(os.path.getsize(full_name)), full_name]) + +def walk_all_subdirectories(path, output_file_name): + # count = 0 + with open(output_file_name, "w") as csvfile: + csvwriter = csv.writer(csvfile, delimiter=':', quotechar='|', quoting=csv.QUOTE_MINIMAL) + for current_dir, sub_dirs, files in os.walk(path): + print(current_dir) + process_directory_csv(current_dir, sub_dirs, files, csvwriter) + csvfile.flush() + # DEBUG CODE - process only 5 directories + # count += 1 + # if count >= 10: + # csvfile.close() + # break; + csvfile.close() + +def sort_file(inname, outname): + input_file = open(inname, "r") + output_file = open(outname, "w", 1) + lines = [] # give lines variable a type of list + for line in input_file: + lines.append(line) + lines.sort() + for line in lines: + output_file.write(line) + input_file.close() + output_file.close() + +def generate_delete_commands(sortedfile, outname): + import csv + output_file = open(outname, "w", 1) + previous_checksum = "IMPOSSIBLE_CHECKSUM" + with open(sortedfile) as f: + reader = csv.reader(f, delimiter=':', quotechar='|', quoting=csv.QUOTE_MINIMAL) + for row in reader: + # print(row[0], row) + if previous_checksum == row[0]: + output_file.write("rm '" + row[2] + "'\n") + print("removing " + row[2]) + os.remove(row[2]) + previous_checksum = row[0] + f.close() + output_file.close() + +# Main program follows + +directory_name = ".." +unsorted_file_name = "filelist.csv" +sorted_file_name = "sortedfilelist.csv" +delete_command_file_name = "deletecommands.sh" + +if __name__ == '__main__': + walk_all_subdirectories('..', unsorted_file_name) + sort_file(unsorted_file_name, sorted_file_name) + generate_delete_commands(sorted_file_name, delete_command_file_name) +
99395e345f74bbedd29fd45eebe0738a3b5f4729
ckanext/archiver/tests/test_api.py
ckanext/archiver/tests/test_api.py
import pytest import tempfile from ckan import model from ckan import plugins from ckan.tests import factories import ckan.tests.helpers as helpers from ckanext.archiver import model as archiver_model from ckanext.archiver.tasks import update_package @pytest.mark.usefixtures('with_plugins') @pytest.mark.ckan_config("ckanext-archiver.cache_url_root", "http://localhost:50001/resources/") @pytest.mark.ckan_config("ckanext-archiver.max_content_length", 1000000) @pytest.mark.ckan_config("ckan.plugins", "archiver testipipe") class TestApi(object): @pytest.fixture(autouse=True) @pytest.mark.usefixtures(u"clean_db") def initial_data(cls, clean_db): archiver_model.init_tables(model.meta.engine) cls.temp_dir = tempfile.mkdtemp() def test_package_show(self, client): url = client + '/?status=200&content=test&content-type=csv' testipipe = plugins.get_plugin('testipipe') testipipe.reset() pkg_dict = { 'name': 'test-package-api', 'resources': [ { 'url': url, 'format': 'TXT', 'description': 'Test' } ] } pkg = factories.Dataset(**pkg_dict) update_package(pkg['id']) result = helpers.call_action( "package_show", id=pkg["id"] ) print(result) assert 'archiver' in result.keys()
Test api endpoint for package show
Test api endpoint for package show
Python
mit
ckan/ckanext-archiver,ckan/ckanext-archiver,ckan/ckanext-archiver
--- +++ @@ -0,0 +1,48 @@ +import pytest +import tempfile + +from ckan import model +from ckan import plugins +from ckan.tests import factories +import ckan.tests.helpers as helpers + +from ckanext.archiver import model as archiver_model +from ckanext.archiver.tasks import update_package + + +@pytest.mark.usefixtures('with_plugins') +@pytest.mark.ckan_config("ckanext-archiver.cache_url_root", "http://localhost:50001/resources/") +@pytest.mark.ckan_config("ckanext-archiver.max_content_length", 1000000) +@pytest.mark.ckan_config("ckan.plugins", "archiver testipipe") +class TestApi(object): + + @pytest.fixture(autouse=True) + @pytest.mark.usefixtures(u"clean_db") + def initial_data(cls, clean_db): + archiver_model.init_tables(model.meta.engine) + cls.temp_dir = tempfile.mkdtemp() + + def test_package_show(self, client): + url = client + '/?status=200&content=test&content-type=csv' + testipipe = plugins.get_plugin('testipipe') + testipipe.reset() + + pkg_dict = { + 'name': 'test-package-api', + 'resources': [ + { + 'url': url, + 'format': 'TXT', + 'description': 'Test' + } + ] + } + pkg = factories.Dataset(**pkg_dict) + update_package(pkg['id']) + + result = helpers.call_action( + "package_show", + id=pkg["id"] + ) + print(result) + assert 'archiver' in result.keys()
797249c42c8c1c0d6eda05dbf9e9d16d2706b373
h2o-py/tests/testdir_algos/deepwater/pyunit_lenet_deepwater.py
h2o-py/tests/testdir_algos/deepwater/pyunit_lenet_deepwater.py
from __future__ import print_function import sys, os sys.path.insert(1, os.path.join("..","..","..")) import h2o from tests import pyunit_utils from h2o.estimators.deepwater import H2ODeepWaterEstimator def deepwater_lenet(): print("Test checks if Deep Water works fine with a multiclass image dataset") frame = h2o.import_file(pyunit_utils.locate("bigdata/laptop/deepwater/imagenet/cat_dog_mouse.csv")) print(frame.head(5)) model = H2ODeepWaterEstimator(epochs=100, rate=1e-3, network='lenet', score_interval=0, train_samples_per_iteration=1000) model.train(x=[0],y=1, training_frame=frame) model.show() error = model.model_performance(train=True).mean_per_class_error() assert error < 0.1, "mean classification error is too high : " + str(error) if __name__ == "__main__": pyunit_utils.standalone_test(deepwater_lenet) else: deepwater_lenet()
Add LeNet example with custom scoring and train_samples_per_iteration.
Add LeNet example with custom scoring and train_samples_per_iteration.
Python
apache-2.0
mathemage/h2o-3,h2oai/h2o-3,spennihana/h2o-3,spennihana/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,michalkurka/h2o-3,spennihana/h2o-3,jangorecki/h2o-3,michalkurka/h2o-3,mathemage/h2o-3,h2oai/h2o-dev,mathemage/h2o-3,h2oai/h2o-3,mathemage/h2o-3,h2oai/h2o-dev,h2oai/h2o-dev,michalkurka/h2o-3,jangorecki/h2o-3,michalkurka/h2o-3,spennihana/h2o-3,spennihana/h2o-3,h2oai/h2o-dev,jangorecki/h2o-3,michalkurka/h2o-3,mathemage/h2o-3,spennihana/h2o-3,jangorecki/h2o-3,h2oai/h2o-dev,mathemage/h2o-3,michalkurka/h2o-3,h2oai/h2o-3,jangorecki/h2o-3,h2oai/h2o-dev,mathemage/h2o-3,h2oai/h2o-3,h2oai/h2o-dev,jangorecki/h2o-3,spennihana/h2o-3,h2oai/h2o-3,jangorecki/h2o-3,h2oai/h2o-3,h2oai/h2o-3
--- +++ @@ -0,0 +1,22 @@ +from __future__ import print_function +import sys, os +sys.path.insert(1, os.path.join("..","..","..")) +import h2o +from tests import pyunit_utils +from h2o.estimators.deepwater import H2ODeepWaterEstimator + +def deepwater_lenet(): + print("Test checks if Deep Water works fine with a multiclass image dataset") + + frame = h2o.import_file(pyunit_utils.locate("bigdata/laptop/deepwater/imagenet/cat_dog_mouse.csv")) + print(frame.head(5)) + model = H2ODeepWaterEstimator(epochs=100, rate=1e-3, network='lenet', score_interval=0, train_samples_per_iteration=1000) + model.train(x=[0],y=1, training_frame=frame) + model.show() + error = model.model_performance(train=True).mean_per_class_error() + assert error < 0.1, "mean classification error is too high : " + str(error) + +if __name__ == "__main__": + pyunit_utils.standalone_test(deepwater_lenet) +else: + deepwater_lenet()
8cdbbbaf33cd09bc742761ce8cd5b79b185710cd
webtool/server/management/commands/timer_update.py
webtool/server/management/commands/timer_update.py
# -*- coding: utf-8 -*- from __future__ import absolute_import from __future__ import unicode_literals from __future__ import print_function import io import datetime from django.core.management.base import BaseCommand from server.models import Instruction, Tour, Talk, Session, Season from server.views.bulletin import Activities from server.views.bulletin.translator import Translator class Command(BaseCommand): leave_locale_alone = True help = 'Update activities regarding the current date' def __init__(self, *args, **kwargs): super(Command, self).__init__(*args, **kwargs) def handle(self, *args, **options): season = Season.objects.get(current=True) canceled = season.state_list.get(name='Ausgefallen') completed = season.state_list.get(name='Durchgeführt') not_touch = (canceled.id, completed.id) today = datetime.date.today() for instruction in Instruction.objects.filter(topic__seasons=season).exclude(state_id__in=not_touch): event = instruction.instruction event_done = ((event.end_date is None and event.start_date < today) or (event.end_date and event.end_date < today)) if event_done: instruction.state = completed instruction.save() instruction.instruction.save() for event in instruction.meeting_list.all(): event.save() for tour in Tour.objects.filter(season=season).exclude(state_id__in=not_touch): event = tour.tour event_done = ((event.end_date is None and event.start_date < today) or (event.end_date and event.end_date < today)) if event_done: tour.state = completed tour.save() tour.deadline.save() if tour.preliminary: tour.preliminary.save() tour.tour.save() for talk in Talk.objects.filter(season=season).exclude(state_id__in=not_touch): event = talk.talk event_done = ((event.end_date is None and event.start_date < today) or (event.end_date and event.end_date < today)) if event_done: talk.state = completed talk.save() talk.talk.save() for session in Session.objects.filter(collective__seasons=season).exclude(state_id__in=not_touch): event = session.session event_done = ((event.end_date is None and event.start_date < today) or (event.end_date and event.end_date < today)) if event_done: session.state = completed session.save() session.session.save()
Introduce a timer based update of activities
Introduce a timer based update of activities
Python
bsd-2-clause
wodo/WebTool3,wodo/WebTool3,wodo/WebTool3,wodo/WebTool3
--- +++ @@ -0,0 +1,72 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import +from __future__ import unicode_literals +from __future__ import print_function + +import io + +import datetime +from django.core.management.base import BaseCommand + +from server.models import Instruction, Tour, Talk, Session, Season +from server.views.bulletin import Activities +from server.views.bulletin.translator import Translator + + +class Command(BaseCommand): + leave_locale_alone = True + help = 'Update activities regarding the current date' + + def __init__(self, *args, **kwargs): + super(Command, self).__init__(*args, **kwargs) + + def handle(self, *args, **options): + + season = Season.objects.get(current=True) + + canceled = season.state_list.get(name='Ausgefallen') + completed = season.state_list.get(name='Durchgeführt') + not_touch = (canceled.id, completed.id) + + today = datetime.date.today() + + for instruction in Instruction.objects.filter(topic__seasons=season).exclude(state_id__in=not_touch): + event = instruction.instruction + event_done = ((event.end_date is None and event.start_date < today) or + (event.end_date and event.end_date < today)) + if event_done: + instruction.state = completed + instruction.save() + instruction.instruction.save() + for event in instruction.meeting_list.all(): + event.save() + + for tour in Tour.objects.filter(season=season).exclude(state_id__in=not_touch): + event = tour.tour + event_done = ((event.end_date is None and event.start_date < today) or + (event.end_date and event.end_date < today)) + if event_done: + tour.state = completed + tour.save() + tour.deadline.save() + if tour.preliminary: + tour.preliminary.save() + tour.tour.save() + + for talk in Talk.objects.filter(season=season).exclude(state_id__in=not_touch): + event = talk.talk + event_done = ((event.end_date is None and event.start_date < today) or + (event.end_date and event.end_date < today)) + if event_done: + talk.state = completed + talk.save() + talk.talk.save() + + for session in Session.objects.filter(collective__seasons=season).exclude(state_id__in=not_touch): + event = session.session + event_done = ((event.end_date is None and event.start_date < today) or + (event.end_date and event.end_date < today)) + if event_done: + session.state = completed + session.save() + session.session.save()
6693172856655329d99f038d54b1d8819fc1a9b6
scripts/examples/02-Board-Control/native_emitters.py
scripts/examples/02-Board-Control/native_emitters.py
import time @micropython.asm_thumb def asm(): movw(r0, 42) @micropython.viper def viper(a, b): return a + b @micropython.native def native(a, b): return a + b print(asm()) print(viper(1, 2)) print(native(1, 2))
Add native code emitters example.
Add native code emitters example.
Python
mit
kwagyeman/openmv,openmv/openmv,openmv/openmv,iabdalkader/openmv,openmv/openmv,openmv/openmv,iabdalkader/openmv,iabdalkader/openmv,kwagyeman/openmv,kwagyeman/openmv,iabdalkader/openmv,kwagyeman/openmv
--- +++ @@ -0,0 +1,19 @@ +import time + +@micropython.asm_thumb +def asm(): + movw(r0, 42) + +@micropython.viper +def viper(a, b): + return a + b + +@micropython.native +def native(a, b): + return a + b + + +print(asm()) +print(viper(1, 2)) +print(native(1, 2)) +
fa6060a21767a0b5b2b3a10e4301e0c1a30134cb
i8c/tests/test_opt_lit0_cmp_before_bra.py
i8c/tests/test_opt_lit0_cmp_before_bra.py
from i8c.tests import TestCase SOURCE1 = """\ define test::optimize_cmp_bra_const_const returns ptr argument ptr x dup load NULL beq return_the_null deref ptr return return_the_null: """ SOURCE2 = """\ define test::optimize_cmp_bra_const_const returns ptr argument ptr x dup load NULL bne dereference return dereference: deref ptr """ class TestOptimizeLit0CmpBeforeBra(TestCase): def test_optimize_lit0_cmp_before_bra(self): """Check that lit0,cmp before bra is eliminated.""" for source in SOURCE1, SOURCE2: tree, output = self.compile(source) self.assertEqual(["dup", "bra", "skip", "deref"], output.opnames)
Test the lit0,cmp before bra eliminator
Test the lit0,cmp before bra eliminator
Python
lgpl-2.1
gbenson/i8c
--- +++ @@ -0,0 +1,32 @@ +from i8c.tests import TestCase + +SOURCE1 = """\ +define test::optimize_cmp_bra_const_const returns ptr + argument ptr x + + dup + load NULL + beq return_the_null + deref ptr + return +return_the_null: +""" + +SOURCE2 = """\ +define test::optimize_cmp_bra_const_const returns ptr + argument ptr x + + dup + load NULL + bne dereference + return +dereference: + deref ptr +""" + +class TestOptimizeLit0CmpBeforeBra(TestCase): + def test_optimize_lit0_cmp_before_bra(self): + """Check that lit0,cmp before bra is eliminated.""" + for source in SOURCE1, SOURCE2: + tree, output = self.compile(source) + self.assertEqual(["dup", "bra", "skip", "deref"], output.opnames)
a3a48824b36ef62edaf128379f1baec5482166e7
src/nodeconductor_saltstack/migrations/0005_resource_error_message.py
src/nodeconductor_saltstack/migrations/0005_resource_error_message.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('nodeconductor_saltstack', '0004_remove_useless_spl_fields'), ] operations = [ migrations.AddField( model_name='domain', name='error_message', field=models.TextField(blank=True), preserve_default=True, ), migrations.AddField( model_name='site', name='error_message', field=models.TextField(blank=True), preserve_default=True, ), ]
Save error_message for resources (SAAS-982)
Save error_message for resources (SAAS-982)
Python
mit
opennode/nodeconductor-saltstack
--- +++ @@ -0,0 +1,26 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import models, migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ('nodeconductor_saltstack', '0004_remove_useless_spl_fields'), + ] + + operations = [ + migrations.AddField( + model_name='domain', + name='error_message', + field=models.TextField(blank=True), + preserve_default=True, + ), + migrations.AddField( + model_name='site', + name='error_message', + field=models.TextField(blank=True), + preserve_default=True, + ), + ]
779fb015913a17fcb8fb290515845e6b47c3ae50
latex2markdown.py
latex2markdown.py
""" A Very simple tool to convert latex documents to markdown documents """ import re span_substitutions = [ (r'\\emph\{(.+)\}', r'*\1*'), (r'\\textbf\{(.+)\}', r'**\1**'), (r'\\verb;(.+);', r'`\1`'), (r'\\includegraphics\{(.+)\}', r'![](\1)'), ] def convert_span_elements(line): """ Converts all recognizable span elements into markdown """ for (f, r) in span_substitutions: p = re.compile(f) line = p.sub(r, line) return line # This next bit is to test the conversion as it builds from sys import stdin if __name__=="__main__": for line in stdin: print(convert_span_elements(line),end='')
Create the converter (with span-conversion functionality)
Create the converter (with span-conversion functionality)
Python
mit
jladan/latex2markdown
--- +++ @@ -0,0 +1,27 @@ +""" +A Very simple tool to convert latex documents to markdown documents +""" +import re + + +span_substitutions = [ + (r'\\emph\{(.+)\}', r'*\1*'), + (r'\\textbf\{(.+)\}', r'**\1**'), + (r'\\verb;(.+);', r'`\1`'), + (r'\\includegraphics\{(.+)\}', r'![](\1)'), + ] + +def convert_span_elements(line): + """ Converts all recognizable span elements into markdown + """ + for (f, r) in span_substitutions: + p = re.compile(f) + line = p.sub(r, line) + return line + +# This next bit is to test the conversion as it builds +from sys import stdin +if __name__=="__main__": + for line in stdin: + print(convert_span_elements(line),end='') +
6014dab06ed2275c5703ab9f9e63272656733c69
moj_utils/rest.py
moj_utils/rest.py
from django.conf import settings def retrieve_all_pages(api_endpoint, **kwargs): """ Some MTP apis are paginated, this method loads all pages into a single results list :param api_endpoint: slumber callable, e.g. `[api_client].cashbook.transactions.locked.get` :param kwargs: additional arguments to pass into api callable """ loaded_results = [] offset = 0 while True: response = api_endpoint(limit=settings.REQUEST_PAGE_SIZE, offset=offset, **kwargs) count = response.get('count', 0) loaded_results += response.get('results', []) if len(loaded_results) >= count: break offset += settings.REQUEST_PAGE_SIZE return loaded_results
Add retrieve_all_pages util method from mtp-cashbook
Add retrieve_all_pages util method from mtp-cashbook This method is duplicated in functionality in multiple front end apps.
Python
mit
ministryofjustice/django-utils,ministryofjustice/django-utils
--- +++ @@ -0,0 +1,22 @@ +from django.conf import settings + + +def retrieve_all_pages(api_endpoint, **kwargs): + """ + Some MTP apis are paginated, this method loads all pages into a single results list + :param api_endpoint: slumber callable, e.g. `[api_client].cashbook.transactions.locked.get` + :param kwargs: additional arguments to pass into api callable + """ + loaded_results = [] + + offset = 0 + while True: + response = api_endpoint(limit=settings.REQUEST_PAGE_SIZE, offset=offset, + **kwargs) + count = response.get('count', 0) + loaded_results += response.get('results', []) + if len(loaded_results) >= count: + break + offset += settings.REQUEST_PAGE_SIZE + + return loaded_results
c675fe2a82733ef210bf287df277f8ae956a4295
rarbg-get.py
rarbg-get.py
#!env /usr/bin/python3 import sys import urllib.parse import urllib.request def main(): search = sys.argv[1] url = 'http://rarbg.to/torrents.php?order=seeders&by=DESC&search=' url = url + search print(url) req = urllib.request.Request(url, headers={'User-Agent' : "Magic Browser"}) resp = urllib.request.urlopen(req) respData = resp.read() if __name__ == '__main__': main()
Add beginning of main script
Add beginning of main script
Python
mit
jadams/rarbg-get
--- +++ @@ -0,0 +1,17 @@ +#!env /usr/bin/python3 + +import sys +import urllib.parse +import urllib.request + +def main(): + search = sys.argv[1] + url = 'http://rarbg.to/torrents.php?order=seeders&by=DESC&search=' + url = url + search + print(url) + req = urllib.request.Request(url, headers={'User-Agent' : "Magic Browser"}) + resp = urllib.request.urlopen(req) + respData = resp.read() + +if __name__ == '__main__': + main()
d2bcba204d36a8ffd1e6a1ed79b89fcb6f1c88c5
ideas/test_kmc.py
ideas/test_kmc.py
# This code will test out the idea of using kmc to # 1. quickly enumerate the k-mers # 2. intersect these with the training database, output as fasta # 3. use that reduced fasta of intersecting kmers as the query to CMash #################################################################### # First, I will need to dump the training database to a fasta file from CMash import MinHash as MH import os import blist training_out_file = '/nfs1/Koslicki_Lab/koslickd/KMC_test/NathanRefSeqTraining60mers.fa' training_data ='/nfs1/Koslicki_Lab/koslickd/MiCOPCMash/TrainingData/NathanRefSeq/micopdb_n_1000_k_60.h5' training_file_names = "/nfs1/Koslicki_Lab/koslickd/MiCOPCMash/TrainingData/NathanRefSeq/absolute_file_names.txt" file_names = [] with open(training_file_names, 'r') as fid: iter = 0 for line in fid.readlines(): line = line.strip() file_names.append(os.path.basename(line)) iter += 1 if iter > 1000: break all_kmers = blist.blist() for file_name in file_names: sketch = MH.import_multiple_from_single_hdf5(training_data, import_list=[file_name])[0] all_kmers += sketch._kmers all_kmers_set = set(all_kmers) with open(training_out_file, 'w') as fid: iter = 0 for kmer in all_kmers_set: fid.write(">seq_%d\n" % iter) fid.write("%s\n" % kmer) iter += 1 ##########################################################################
Add file to test out kmc approach. Dump training k-mers to fasta file
Add file to test out kmc approach. Dump training k-mers to fasta file
Python
bsd-3-clause
dkoslicki/CMash,dkoslicki/CMash
--- +++ @@ -0,0 +1,41 @@ +# This code will test out the idea of using kmc to +# 1. quickly enumerate the k-mers +# 2. intersect these with the training database, output as fasta +# 3. use that reduced fasta of intersecting kmers as the query to CMash + +#################################################################### +# First, I will need to dump the training database to a fasta file +from CMash import MinHash as MH +import os +import blist + +training_out_file = '/nfs1/Koslicki_Lab/koslickd/KMC_test/NathanRefSeqTraining60mers.fa' +training_data ='/nfs1/Koslicki_Lab/koslickd/MiCOPCMash/TrainingData/NathanRefSeq/micopdb_n_1000_k_60.h5' +training_file_names = "/nfs1/Koslicki_Lab/koslickd/MiCOPCMash/TrainingData/NathanRefSeq/absolute_file_names.txt" + +file_names = [] +with open(training_file_names, 'r') as fid: + iter = 0 + for line in fid.readlines(): + line = line.strip() + file_names.append(os.path.basename(line)) + iter += 1 + if iter > 1000: + break + +all_kmers = blist.blist() +for file_name in file_names: + sketch = MH.import_multiple_from_single_hdf5(training_data, import_list=[file_name])[0] + all_kmers += sketch._kmers + +all_kmers_set = set(all_kmers) + +with open(training_out_file, 'w') as fid: + iter = 0 + for kmer in all_kmers_set: + fid.write(">seq_%d\n" % iter) + fid.write("%s\n" % kmer) + iter += 1 + +########################################################################## +
02ad029840b2e770bc802fd7f8504498cb0f756d
lib/ansible/plugins/test/mathstuff.py
lib/ansible/plugins/test/mathstuff.py
# (c) 2016, Ansible, Inc # # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. __metaclass__ = type def issubset(a, b): return set(a) <= set(b) def issuperset(a, b): return set(a) >= set(b) class TestModule(object): ''' Ansible math jinja2 tests ''' def tests(self): return { # set theory 'issubset': issubset, 'issuperset': issuperset, }
Add `issubset` and `issuperset` tests
Add `issubset` and `issuperset` tests
Python
mit
thaim/ansible,thaim/ansible
--- +++ @@ -0,0 +1,34 @@ +# (c) 2016, Ansible, Inc +# +# This file is part of Ansible +# +# Ansible is free software: you can redistribute it and/or modify +# it under the terms of the GNU General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# Ansible is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with Ansible. If not, see <http://www.gnu.org/licenses/>. + +__metaclass__ = type + +def issubset(a, b): + return set(a) <= set(b) + +def issuperset(a, b): + return set(a) >= set(b) + +class TestModule(object): + ''' Ansible math jinja2 tests ''' + + def tests(self): + return { + # set theory + 'issubset': issubset, + 'issuperset': issuperset, + }
bf86584829f56f91b363f251d77f3157f952db0f
tests/test_cyprep.py
tests/test_cyprep.py
import unittest import numpy as np import yatsm._cyprep class TestCyPrep(unittest.TestCase): @classmethod def setUpClass(cls): # Test data n_band = 7 n_mask = 50 n_images = 1000 cls.data = np.random.randint( 0, 10000, size=(n_band, n_images)).astype(np.int32) for b in range(n_band): cls.data[b, np.random.choice(np.arange(0, n_images), size=n_mask, replace=False)] = 16000 cls.mins = np.repeat(0, n_band).astype(np.int16) cls.maxes = np.repeat(10000, n_band).astype(np.int16) def test_get_valid_mask(self): truth = np.all([((b > _min) & (b < _max)) for b, _min, _max in zip(np.rollaxis(self.data, 0), self.mins, self.maxes)], axis=0) np.testing.assert_equal( truth, yatsm._cyprep.get_valid_mask(self.data, self.mins, self.maxes)) if __name__ == '__main__': unittest.main()
Add tests for masking of data based on being within a range of values
Add tests for masking of data based on being within a range of values
Python
mit
ceholden/yatsm,c11/yatsm,jmorton/yatsm,valpasq/yatsm,ceholden/yatsm,valpasq/yatsm,jmorton/yatsm,c11/yatsm,jmorton/yatsm
--- +++ @@ -0,0 +1,38 @@ +import unittest + +import numpy as np + +import yatsm._cyprep + + +class TestCyPrep(unittest.TestCase): + + @classmethod + def setUpClass(cls): + # Test data + n_band = 7 + n_mask = 50 + n_images = 1000 + + cls.data = np.random.randint( + 0, 10000, size=(n_band, n_images)).astype(np.int32) + for b in range(n_band): + cls.data[b, np.random.choice(np.arange(0, n_images), + size=n_mask, replace=False)] = 16000 + + cls.mins = np.repeat(0, n_band).astype(np.int16) + cls.maxes = np.repeat(10000, n_band).astype(np.int16) + + def test_get_valid_mask(self): + truth = np.all([((b > _min) & (b < _max)) for b, _min, _max in + zip(np.rollaxis(self.data, 0), + self.mins, + self.maxes)], axis=0) + + np.testing.assert_equal( + truth, + yatsm._cyprep.get_valid_mask(self.data, self.mins, self.maxes)) + + +if __name__ == '__main__': + unittest.main()
aa78a2670766b0a5e093a1876cb402ed513573bd
openfisca_france/scripts/parameters/explore_parameters_unit.py
openfisca_france/scripts/parameters/explore_parameters_unit.py
# -*- coding: utf-8 -*- from openfisca_core.parameters import ParameterNode, Scale from openfisca_france import FranceTaxBenefitSystem tax_benefit_system = FranceTaxBenefitSystem() parameters = tax_benefit_system.parameters def get_parameters_by_unit(parameter, parameters_by_unit = None): if parameters_by_unit is None: parameters_by_unit = dict( scale = list(), none = list(), currency = list(), rate = list(), year = list(), ) for name, sub_parameter in parameter.children.items(): if isinstance(sub_parameter, ParameterNode): get_parameters_by_unit(sub_parameter, parameters_by_unit) else: if isinstance(sub_parameter, Scale): parameters_by_unit['scale'].append(sub_parameter) elif sub_parameter.unit is None: parameters_by_unit['none'].append(sub_parameter) elif sub_parameter.unit == "/1": parameters_by_unit['rate'].append(sub_parameter) elif sub_parameter.unit == "currency": parameters_by_unit['currency'].append(sub_parameter) elif sub_parameter.unit == "year": parameters_by_unit['year'].append(sub_parameter) else: raise ValueError("Parameter {} has a stange unit {}".format( sub_parameter.name, sub_parameter.unit)) return parameters_by_unit if __name__ == '__main__': parameters_by_unit = get_parameters_by_unit(parameters) print('Distribution of parameters types:') for type_, sub_parameters in parameters_by_unit.items(): print(type_, len(parameters_by_unit[type_])) print('\n') print('List of parameters with no units') for param in parameters_by_unit['none']: print (param.name)
Add script to explore parameters units
Add script to explore parameters units
Python
agpl-3.0
antoinearnoud/openfisca-france,antoinearnoud/openfisca-france,sgmap/openfisca-france,sgmap/openfisca-france
--- +++ @@ -0,0 +1,53 @@ +# -*- coding: utf-8 -*- + + +from openfisca_core.parameters import ParameterNode, Scale +from openfisca_france import FranceTaxBenefitSystem + + +tax_benefit_system = FranceTaxBenefitSystem() +parameters = tax_benefit_system.parameters + + +def get_parameters_by_unit(parameter, parameters_by_unit = None): + if parameters_by_unit is None: + parameters_by_unit = dict( + scale = list(), + none = list(), + currency = list(), + rate = list(), + year = list(), + ) + for name, sub_parameter in parameter.children.items(): + if isinstance(sub_parameter, ParameterNode): + get_parameters_by_unit(sub_parameter, parameters_by_unit) + else: + if isinstance(sub_parameter, Scale): + parameters_by_unit['scale'].append(sub_parameter) + elif sub_parameter.unit is None: + parameters_by_unit['none'].append(sub_parameter) + elif sub_parameter.unit == "/1": + parameters_by_unit['rate'].append(sub_parameter) + elif sub_parameter.unit == "currency": + parameters_by_unit['currency'].append(sub_parameter) + elif sub_parameter.unit == "year": + parameters_by_unit['year'].append(sub_parameter) + else: + raise ValueError("Parameter {} has a stange unit {}".format( + sub_parameter.name, sub_parameter.unit)) + + return parameters_by_unit + + +if __name__ == '__main__': + parameters_by_unit = get_parameters_by_unit(parameters) + print('Distribution of parameters types:') + for type_, sub_parameters in parameters_by_unit.items(): + print(type_, len(parameters_by_unit[type_])) + + print('\n') + print('List of parameters with no units') + + for param in parameters_by_unit['none']: + print (param.name) +
26ab37868e67b5b815cf8df67cc04876ff44c148
tests/rules_tests/isValid_tests/NongrammarEntitiesTest.py
tests/rules_tests/isValid_tests/NongrammarEntitiesTest.py
#!/usr/bin/env python """ :Author Patrik Valkovic :Created 23.06.2017 16:39 :Licence GNUv3 Part of grammpy """ from unittest import main, TestCase from grammpy import Rule from .grammar import * class NongrammarEntitiesTest(TestCase): pass if __name__ == '__main__': main()
Add file for Nongrammar entities tests
Add file for Nongrammar entities tests
Python
mit
PatrikValkovic/grammpy
--- +++ @@ -0,0 +1,20 @@ +#!/usr/bin/env python +""" +:Author Patrik Valkovic +:Created 23.06.2017 16:39 +:Licence GNUv3 +Part of grammpy + +""" + +from unittest import main, TestCase +from grammpy import Rule +from .grammar import * + + +class NongrammarEntitiesTest(TestCase): + pass + + +if __name__ == '__main__': + main()
fa3a02e6660ce556defc2f2c6008c6eb24eb71c1
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Sampler.py
Sketches/JT/Jam/library/trunk/Kamaelia/Apps/Jam/Audio/Sampler.py
import time import wave import pygame import numpy import Axon from Axon.SchedulingComponent import SchedulingComponent class WavVoice(SchedulingComponent): bufferSize = 1024 def __init__(self, fileName, **argd): super(WavVoice, self).__init__(**argd) self.on = False self.wavFile = wave.open(fileName) self.sampleRate = self.wavFile.getframerate() self.period = float(self.bufferSize)/self.sampleRate self.frame = 0 self.lastSendTime = time.time() self.scheduleAbs("Send", self.lastSendTime + self.period) def main(self): while 1: if self.dataReady("inbox"): address, arguments = self.recv("inbox") address = address.split("/")[-1] if address == "On": self.on = True self.wavFile.rewind() self.frame = 0 if address == "Off": self.on = False if self.dataReady("event"): self.recv("event") if self.on: if self.frame < self.wavFile.getnframes(): sample = self.wavFile.readframes(self.bufferSize) sample = numpy.frombuffer(sample, dtype="int16") self.frame += len(sample) if len(sample) < self.bufferSize: # Pad with zeroes padSize = self.bufferSize - len(sample) sample = numpy.append(sample, numpy.zeros(padSize)) # Convert to float sample = sample.astype("float64") # Scale to -1 - 1 sample /= 2**(8 * self.wavFile.getsampwidth() - 1) else: sample = numpy.zeros(self.bufferSize) else: sample = numpy.zeros(self.bufferSize) self.send(sample, "outbox") self.lastSendTime += self.period self.scheduleAbs("Send", self.lastSendTime + self.period) if not self.anyReady(): self.pause() if __name__ == "__main__": from Kamaelia.Chassis.Pipeline import Pipeline from Kamaelia.Apps.Jam.Util.Numpy import TypeConverter from Kamaelia.Codec.Vorbis import AOAudioPlaybackAdaptor from Kamaelia.Util.PureTransformer import PureTransformer from Kamaelia.Apps.Jam.UI.StepSequencer import StepSequencer from Kamaelia.Apps.Jam.Audio.Synth import Synth from Kamaelia.Apps.Jam.Audio.Polyphony import Targetter files = ["Ride", "HH", "Snare", "Kick"] files = ["/home/joe/Desktop/%s.wav"%fileName for fileName in files] def voiceGenerator(): for i in range(4): yield WavVoice(files[i]) Pipeline(StepSequencer(stepsPerBeat=4), Synth(voiceGenerator, polyphoniser=Targetter, polyphony=4), PureTransformer(lambda x:x*(2**15-1)), TypeConverter(type="int16"), AOAudioPlaybackAdaptor()).run()
Add a simple sampler for playing wav files triggered by note on messages
Add a simple sampler for playing wav files triggered by note on messages
Python
apache-2.0
sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia,sparkslabs/kamaelia
--- +++ @@ -0,0 +1,80 @@ +import time +import wave +import pygame +import numpy +import Axon +from Axon.SchedulingComponent import SchedulingComponent + +class WavVoice(SchedulingComponent): + bufferSize = 1024 + def __init__(self, fileName, **argd): + super(WavVoice, self).__init__(**argd) + + self.on = False + + self.wavFile = wave.open(fileName) + self.sampleRate = self.wavFile.getframerate() + self.period = float(self.bufferSize)/self.sampleRate + + self.frame = 0 + + self.lastSendTime = time.time() + self.scheduleAbs("Send", self.lastSendTime + self.period) + + def main(self): + while 1: + if self.dataReady("inbox"): + address, arguments = self.recv("inbox") + address = address.split("/")[-1] + if address == "On": + self.on = True + self.wavFile.rewind() + self.frame = 0 + if address == "Off": + self.on = False + + if self.dataReady("event"): + self.recv("event") + if self.on: + if self.frame < self.wavFile.getnframes(): + sample = self.wavFile.readframes(self.bufferSize) + sample = numpy.frombuffer(sample, dtype="int16") + self.frame += len(sample) + if len(sample) < self.bufferSize: + # Pad with zeroes + padSize = self.bufferSize - len(sample) + sample = numpy.append(sample, numpy.zeros(padSize)) + # Convert to float + sample = sample.astype("float64") + # Scale to -1 - 1 + sample /= 2**(8 * self.wavFile.getsampwidth() - 1) + else: + sample = numpy.zeros(self.bufferSize) + else: + sample = numpy.zeros(self.bufferSize) + self.send(sample, "outbox") + self.lastSendTime += self.period + self.scheduleAbs("Send", self.lastSendTime + self.period) + + if not self.anyReady(): + self.pause() + +if __name__ == "__main__": + from Kamaelia.Chassis.Pipeline import Pipeline + from Kamaelia.Apps.Jam.Util.Numpy import TypeConverter + from Kamaelia.Codec.Vorbis import AOAudioPlaybackAdaptor + from Kamaelia.Util.PureTransformer import PureTransformer + from Kamaelia.Apps.Jam.UI.StepSequencer import StepSequencer + from Kamaelia.Apps.Jam.Audio.Synth import Synth + from Kamaelia.Apps.Jam.Audio.Polyphony import Targetter + + files = ["Ride", "HH", "Snare", "Kick"] + files = ["/home/joe/Desktop/%s.wav"%fileName for fileName in files] + def voiceGenerator(): + for i in range(4): + yield WavVoice(files[i]) + + Pipeline(StepSequencer(stepsPerBeat=4), Synth(voiceGenerator, polyphoniser=Targetter, polyphony=4), PureTransformer(lambda x:x*(2**15-1)), TypeConverter(type="int16"), AOAudioPlaybackAdaptor()).run() + + +
4e3644234fab9cb14a3d511b24bce3ed8a1446e0
tests/scales/test_minor.py
tests/scales/test_minor.py
# Copyright (c) Paul R. Tagliamonte <tag@pault.ag>, 2015 # # Permission is hereby granted, free of charge, to any person obtaining a # copy of this software and associated documentation files (the "Software"), # to deal in the Software without restriction, including without limitation # the rights to use, copy, modify, merge, publish, distribute, sublicense, # and/or sell copies of the Software, and to permit persons to whom the # Software is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL # THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING # FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER # DEALINGS IN THE SOFTWARE. from muse.scales.minor import NaturalMinorScale from muse.tone import Tone def take(it, count): for _ in range(count): yield next(it) SCALE = ['B♭4', 'C4', 'C♯4', 'E♭4', 'F4', 'F♯4', 'G♯4', 'B♭5'] def test_scale_acending_iteratation(): cs = NaturalMinorScale(Tone(100)) # Bb4 series = list(take(cs.acending(), 8)) assert [x._tone_name() for x in series] == SCALE def test_scale_acending_iteratation(): cs = NaturalMinorScale(Tone(1300)) # Bb5 series = list(take(cs.decending(), 8)) assert [x._tone_name() for x in series] == list(reversed(SCALE))
Add in a minor testcase.
Add in a minor testcase.
Python
mit
paultag/python-muse
--- +++ @@ -0,0 +1,41 @@ +# Copyright (c) Paul R. Tagliamonte <tag@pault.ag>, 2015 +# +# Permission is hereby granted, free of charge, to any person obtaining a +# copy of this software and associated documentation files (the "Software"), +# to deal in the Software without restriction, including without limitation +# the rights to use, copy, modify, merge, publish, distribute, sublicense, +# and/or sell copies of the Software, and to permit persons to whom the +# Software is furnished to do so, subject to the following conditions: +# +# The above copyright notice and this permission notice shall be included in +# all copies or substantial portions of the Software. +# +# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL +# THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING +# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER +# DEALINGS IN THE SOFTWARE. + +from muse.scales.minor import NaturalMinorScale +from muse.tone import Tone + +def take(it, count): + for _ in range(count): + yield next(it) + + +SCALE = ['B♭4', 'C4', 'C♯4', 'E♭4', 'F4', 'F♯4', 'G♯4', 'B♭5'] + + +def test_scale_acending_iteratation(): + cs = NaturalMinorScale(Tone(100)) # Bb4 + series = list(take(cs.acending(), 8)) + assert [x._tone_name() for x in series] == SCALE + + +def test_scale_acending_iteratation(): + cs = NaturalMinorScale(Tone(1300)) # Bb5 + series = list(take(cs.decending(), 8)) + assert [x._tone_name() for x in series] == list(reversed(SCALE))
6f8c64ed6f99493811cab54137a1eed44d851260
scripts/GetGroupAndModuleFromClassName.py
scripts/GetGroupAndModuleFromClassName.py
#!/usr/bin/env python """ Given the path to the ITK Source Dir print group and module of a given class for instance, try: ./GetGroupAndModuleFromClassName /path/to/ITK Image """ import sys import os itk_dir = sys.argv[1] cmakefile = os.path.join( itk_dir, 'CMake', 'UseITK.cmake' ) if not os.path.exists( cmakefile ): print 'Error: wrong path' else: class_name = sys.argv[2] path = '' for root, dirs, files in os.walk( os.path.join( itk_dir, 'Modules' ) ): for f in files: if f == 'itk' + class_name + '.h': path = root if len( path ) != 0: # let's extract the Group temp = path.strip( os.path.join( itk_dir, 'Modules' ) ) temp = temp.strip( 'include' ) GroupModule = temp.split( '/' ) print 'Group: ' + GroupModule[ 0 ] print 'Module: ' + GroupModule[ 1 ] else: print 'Error: this class is not part of itk'
Add python script to get group and module given a class name
Add python script to get group and module given a class name
Python
apache-2.0
InsightSoftwareConsortium/ITKExamples,InsightSoftwareConsortium/ITKExamples,InsightSoftwareConsortium/ITKExamples,InsightSoftwareConsortium/ITKExamples,InsightSoftwareConsortium/ITKExamples
--- +++ @@ -0,0 +1,39 @@ +#!/usr/bin/env python + +""" Given the path to the ITK Source Dir +print group and module of a given class + +for instance, try: + + ./GetGroupAndModuleFromClassName /path/to/ITK Image +""" + +import sys +import os + +itk_dir = sys.argv[1] +cmakefile = os.path.join( itk_dir, 'CMake', 'UseITK.cmake' ) + +if not os.path.exists( cmakefile ): + print 'Error: wrong path' +else: + class_name = sys.argv[2] + path = '' + + for root, dirs, files in os.walk( os.path.join( itk_dir, 'Modules' ) ): + for f in files: + if f == 'itk' + class_name + '.h': + path = root + + if len( path ) != 0: + # let's extract the Group + temp = path.strip( os.path.join( itk_dir, 'Modules' ) ) + temp = temp.strip( 'include' ) + + GroupModule = temp.split( '/' ) + print 'Group: ' + GroupModule[ 0 ] + print 'Module: ' + GroupModule[ 1 ] + + else: + print 'Error: this class is not part of itk' +
acc5c52011db4c8edc615ae3e0cad9cea4fe58b8
spreadflow_observer_fs/test/test_source.py
spreadflow_observer_fs/test/test_source.py
# -*- coding: utf-8 -*- # pylint: disable=too-many-public-methods """ Integration tests for spreadflow filesystem observer source. """ from __future__ import absolute_import from __future__ import division from __future__ import unicode_literals import copy from bson import BSON from datetime import datetime from twisted.internet import defer from mock import Mock from testtools import TestCase, run_test_with from testtools.twistedsupport import AsynchronousDeferredRunTest from spreadflow_core.scheduler import Scheduler from spreadflow_delta.test.matchers import MatchesSendDeltaItemInvocation from spreadflow_observer_fs.source import FilesystemObserverSource def _spawnProcess(processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): """ Spawn process method signature. """ class SpreadflowSourceIntegrationTestCase(TestCase): """ Integration tests for spreadflow filesystem observer source. """ @run_test_with(AsynchronousDeferredRunTest) @defer.inlineCallbacks def test_source_process(self): source = FilesystemObserverSource('*.txt', '/some/directory') reactor = Mock() reactor.spawnProcess = Mock(spec=_spawnProcess) scheduler = Mock() scheduler.send = Mock(spec=Scheduler.send) # Attach source to the scheduler. yield source.attach(scheduler, reactor) self.assertEquals(reactor.spawnProcess.call_count, 1) # Simulate a message directed to the source. msg = { 'port': 'default', 'item': { 'type': 'delta', 'date': datetime(2010, 10, 20, 20, 10), 'inserts': ['abcdefg'], 'deletes': ['hiklmno'], 'data': { 'abcdefg': { 'path': '/some/directory/xyz.txt' } } } } matches = MatchesSendDeltaItemInvocation(copy.deepcopy(msg['item']), source) source.peer.dataReceived(BSON.encode(msg)) self.assertEquals(scheduler.send.call_count, 1) self.assertThat(scheduler.send.call_args, matches)
Add basic test for filesystem observer source
Add basic test for filesystem observer source
Python
mit
znerol/spreadflow-observer-fs,spreadflow/spreadflow-observer-fs
--- +++ @@ -0,0 +1,72 @@ +# -*- coding: utf-8 -*- +# pylint: disable=too-many-public-methods + +""" +Integration tests for spreadflow filesystem observer source. +""" + +from __future__ import absolute_import +from __future__ import division +from __future__ import unicode_literals + +import copy + +from bson import BSON +from datetime import datetime +from twisted.internet import defer + +from mock import Mock +from testtools import TestCase, run_test_with +from testtools.twistedsupport import AsynchronousDeferredRunTest + +from spreadflow_core.scheduler import Scheduler +from spreadflow_delta.test.matchers import MatchesSendDeltaItemInvocation + +from spreadflow_observer_fs.source import FilesystemObserverSource + + +def _spawnProcess(processProtocol, executable, args=(), env={}, path=None, uid=None, gid=None, usePTY=0, childFDs=None): + """ + Spawn process method signature. + """ + +class SpreadflowSourceIntegrationTestCase(TestCase): + """ + Integration tests for spreadflow filesystem observer source. + """ + + @run_test_with(AsynchronousDeferredRunTest) + @defer.inlineCallbacks + def test_source_process(self): + source = FilesystemObserverSource('*.txt', '/some/directory') + + reactor = Mock() + reactor.spawnProcess = Mock(spec=_spawnProcess) + + scheduler = Mock() + scheduler.send = Mock(spec=Scheduler.send) + + # Attach source to the scheduler. + yield source.attach(scheduler, reactor) + self.assertEquals(reactor.spawnProcess.call_count, 1) + + # Simulate a message directed to the source. + msg = { + 'port': 'default', + 'item': { + 'type': 'delta', + 'date': datetime(2010, 10, 20, 20, 10), + 'inserts': ['abcdefg'], + 'deletes': ['hiklmno'], + 'data': { + 'abcdefg': { + 'path': '/some/directory/xyz.txt' + } + } + } + } + + matches = MatchesSendDeltaItemInvocation(copy.deepcopy(msg['item']), source) + source.peer.dataReceived(BSON.encode(msg)) + self.assertEquals(scheduler.send.call_count, 1) + self.assertThat(scheduler.send.call_args, matches)
5273a97ab1da4b809573617d3fc01705c322992f
thecut/authorship/tests/test_forms.py
thecut/authorship/tests/test_forms.py
# -*- coding: utf-8 -*- from __future__ import absolute_import, unicode_literals from django.test import TestCase from django import forms from mock import patch from test_app.models import AuthorshipModel from thecut.authorship.factories import UserFactory from thecut.authorship.forms import AuthorshipMixin class AuthorshipModelForm(AuthorshipMixin, forms.ModelForm): class Meta: model = AuthorshipModel fields = [] class DummyUser(object): pass class TestAuthorshipMixin(TestCase): def test_requires_an_extra_argument_on_creating_an_instance(self): self.assertRaises(TypeError, AuthorshipModelForm) def test_sets_user_attribute(self): dummy_user = DummyUser() form = AuthorshipModelForm(user=dummy_user) self.assertEqual(dummy_user, form.user) class DummyUnsavedModel(object): def __init__(self): self.pk = None class TestAuthorshipMixinSave(TestCase): @patch('django.forms.ModelForm.save') def test_calls_super_class_save_method(self, superclass_save): form = AuthorshipModelForm(user=UserFactory()) form.instance = DummyUnsavedModel() form.save() self.assertTrue(superclass_save.called) @patch('django.forms.ModelForm.save') def test_sets_updated_by_to_given_user(self, superclass_save): user = DummyUser() form = AuthorshipModelForm(user=user) form.instance = DummyUnsavedModel() form.cleaned_data = {} form.save() self.assertEqual(user, form.instance.updated_by) @patch('django.forms.ModelForm.save') def test_sets_created_by_if_instance_is_not_saved(self, superclass_save): user = DummyUser() form = AuthorshipModelForm(user=user) form.instance = DummyUnsavedModel() form.cleaned_data = {} form.save() self.assertEqual(user, form.instance.created_by) @patch('django.forms.ModelForm.save') def test_does_not_set_created_by_if_instance_is_saved(self, superclass_save): class DummySavedModel(object): def __init__(self): self.pk = 'arbitrary-value' self.created_by = 'arbitrary-value' user = DummyUser() form = AuthorshipModelForm(user=user) form.instance = DummySavedModel() form.cleaned_data = {} form.save() self.assertNotEqual(user, form.instance.created_by)
Add tests for form mixin.
Add tests for form mixin.
Python
apache-2.0
thecut/thecut-authorship
--- +++ @@ -0,0 +1,94 @@ +# -*- coding: utf-8 -*- +from __future__ import absolute_import, unicode_literals +from django.test import TestCase +from django import forms +from mock import patch +from test_app.models import AuthorshipModel +from thecut.authorship.factories import UserFactory +from thecut.authorship.forms import AuthorshipMixin + + +class AuthorshipModelForm(AuthorshipMixin, forms.ModelForm): + + class Meta: + model = AuthorshipModel + fields = [] + + +class DummyUser(object): + + pass + + +class TestAuthorshipMixin(TestCase): + + def test_requires_an_extra_argument_on_creating_an_instance(self): + self.assertRaises(TypeError, AuthorshipModelForm) + + def test_sets_user_attribute(self): + + dummy_user = DummyUser() + + form = AuthorshipModelForm(user=dummy_user) + + self.assertEqual(dummy_user, form.user) + + +class DummyUnsavedModel(object): + + def __init__(self): + self.pk = None + + +class TestAuthorshipMixinSave(TestCase): + + @patch('django.forms.ModelForm.save') + def test_calls_super_class_save_method(self, superclass_save): + + form = AuthorshipModelForm(user=UserFactory()) + form.instance = DummyUnsavedModel() + + form.save() + + self.assertTrue(superclass_save.called) + + @patch('django.forms.ModelForm.save') + def test_sets_updated_by_to_given_user(self, superclass_save): + user = DummyUser() + form = AuthorshipModelForm(user=user) + form.instance = DummyUnsavedModel() + form.cleaned_data = {} + + form.save() + + self.assertEqual(user, form.instance.updated_by) + + @patch('django.forms.ModelForm.save') + def test_sets_created_by_if_instance_is_not_saved(self, superclass_save): + user = DummyUser() + form = AuthorshipModelForm(user=user) + form.instance = DummyUnsavedModel() + form.cleaned_data = {} + + form.save() + + self.assertEqual(user, form.instance.created_by) + + @patch('django.forms.ModelForm.save') + def test_does_not_set_created_by_if_instance_is_saved(self, + superclass_save): + + class DummySavedModel(object): + + def __init__(self): + self.pk = 'arbitrary-value' + self.created_by = 'arbitrary-value' + + user = DummyUser() + form = AuthorshipModelForm(user=user) + form.instance = DummySavedModel() + form.cleaned_data = {} + + form.save() + + self.assertNotEqual(user, form.instance.created_by)
e838b6d53f131badfbb7b51b4eb268ebb5d7c450
spacy/tests/matcher/test_entity_id.py
spacy/tests/matcher/test_entity_id.py
from __future__ import unicode_literals import spacy from spacy.vocab import Vocab from spacy.matcher import Matcher from spacy.tokens.doc import Doc from spacy.attrs import * import pytest @pytest.fixture def en_vocab(): return spacy.get_lang_class('en').Defaults.create_vocab() def test_init_matcher(en_vocab): matcher = Matcher(en_vocab) assert matcher.n_patterns == 0 assert matcher(Doc(en_vocab, words=[u'Some', u'words'])) == [] def test_add_empty_entity(en_vocab): matcher = Matcher(en_vocab) matcher.add_entity('TestEntity') assert matcher.n_patterns == 0 assert matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) == [] def test_get_entity_attrs(en_vocab): matcher = Matcher(en_vocab) matcher.add_entity('TestEntity') entity = matcher.get_entity('TestEntity') assert entity == {} matcher.add_entity('TestEntity2', attrs={'Hello': 'World'}) entity = matcher.get_entity('TestEntity2') assert entity == {'Hello': 'World'} assert matcher.get_entity('TestEntity') == {} def test_get_entity_via_match(en_vocab): matcher = Matcher(en_vocab) matcher.add_entity('TestEntity', attrs={u'Hello': u'World'}) assert matcher.n_patterns == 0 assert matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) == [] matcher.add_pattern(u'TestEntity', [{ORTH: u'Test'}, {ORTH: u'Entity'}]) assert matcher.n_patterns == 1 matches = matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) assert len(matches) == 1 assert len(matches[0]) == 4 ent_id, label, start, end = matches[0] assert ent_id == matcher.vocab.strings[u'TestEntity'] assert label == 0 assert start == 0 assert end == 2 attrs = matcher.get_entity(ent_id) assert attrs == {u'Hello': u'World'}
Add tests for using the new Entity ID tracking in the rule matcher
Add tests for using the new Entity ID tracking in the rule matcher
Python
mit
banglakit/spaCy,spacy-io/spaCy,recognai/spaCy,explosion/spaCy,raphael0202/spaCy,aikramer2/spaCy,oroszgy/spaCy.hu,Gregory-Howard/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,banglakit/spaCy,raphael0202/spaCy,honnibal/spaCy,banglakit/spaCy,raphael0202/spaCy,recognai/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,banglakit/spaCy,banglakit/spaCy,aikramer2/spaCy,honnibal/spaCy,Gregory-Howard/spaCy,Gregory-Howard/spaCy,raphael0202/spaCy,spacy-io/spaCy,spacy-io/spaCy,recognai/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,banglakit/spaCy,aikramer2/spaCy,spacy-io/spaCy,oroszgy/spaCy.hu,recognai/spaCy,explosion/spaCy,raphael0202/spaCy,recognai/spaCy,oroszgy/spaCy.hu,honnibal/spaCy,explosion/spaCy,explosion/spaCy,explosion/spaCy,oroszgy/spaCy.hu,aikramer2/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,explosion/spaCy,honnibal/spaCy,recognai/spaCy,Gregory-Howard/spaCy
--- +++ @@ -0,0 +1,59 @@ +from __future__ import unicode_literals +import spacy +from spacy.vocab import Vocab +from spacy.matcher import Matcher +from spacy.tokens.doc import Doc +from spacy.attrs import * + +import pytest + + +@pytest.fixture +def en_vocab(): + return spacy.get_lang_class('en').Defaults.create_vocab() + + +def test_init_matcher(en_vocab): + matcher = Matcher(en_vocab) + assert matcher.n_patterns == 0 + assert matcher(Doc(en_vocab, words=[u'Some', u'words'])) == [] + + +def test_add_empty_entity(en_vocab): + matcher = Matcher(en_vocab) + matcher.add_entity('TestEntity') + assert matcher.n_patterns == 0 + assert matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) == [] + + +def test_get_entity_attrs(en_vocab): + matcher = Matcher(en_vocab) + matcher.add_entity('TestEntity') + entity = matcher.get_entity('TestEntity') + assert entity == {} + matcher.add_entity('TestEntity2', attrs={'Hello': 'World'}) + entity = matcher.get_entity('TestEntity2') + assert entity == {'Hello': 'World'} + assert matcher.get_entity('TestEntity') == {} + + +def test_get_entity_via_match(en_vocab): + matcher = Matcher(en_vocab) + matcher.add_entity('TestEntity', attrs={u'Hello': u'World'}) + assert matcher.n_patterns == 0 + assert matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) == [] + matcher.add_pattern(u'TestEntity', [{ORTH: u'Test'}, {ORTH: u'Entity'}]) + assert matcher.n_patterns == 1 + matches = matcher(Doc(en_vocab, words=[u'Test', u'Entity'])) + assert len(matches) == 1 + assert len(matches[0]) == 4 + ent_id, label, start, end = matches[0] + assert ent_id == matcher.vocab.strings[u'TestEntity'] + assert label == 0 + assert start == 0 + assert end == 2 + attrs = matcher.get_entity(ent_id) + assert attrs == {u'Hello': u'World'} + + +
9a6ca54f7cca0bd5f21f0bc590a034e7e3e05b6e
src/icp/apps/user/migrations/0002_add_userprofiles_to_existing_users.py
src/icp/apps/user/migrations/0002_add_userprofiles_to_existing_users.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models import django.contrib.postgres.fields from django.conf import settings def create_user_profiles_for_existing_users(apps, schema_editor): User = apps.get_model('auth', 'User') UserProfile = apps.get_model('user', 'UserProfile') for user in User.objects.all(): UserProfile.objects.create(user=user) class Migration(migrations.Migration): dependencies = [ ('user', '0001_initial') ] operations = [ migrations.RunPython(create_user_profiles_for_existing_users) ]
Add migration to add userprofiles to existing users
Add migration to add userprofiles to existing users All existing users on staging/production will be users of the Pollination Mapper app. Refs #290
Python
apache-2.0
project-icp/bee-pollinator-app,project-icp/bee-pollinator-app,project-icp/bee-pollinator-app,project-icp/bee-pollinator-app
--- +++ @@ -0,0 +1,24 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import migrations, models +import django.contrib.postgres.fields +from django.conf import settings + +def create_user_profiles_for_existing_users(apps, schema_editor): + User = apps.get_model('auth', 'User') + UserProfile = apps.get_model('user', 'UserProfile') + + for user in User.objects.all(): + UserProfile.objects.create(user=user) + + +class Migration(migrations.Migration): + + dependencies = [ + ('user', '0001_initial') + ] + + operations = [ + migrations.RunPython(create_user_profiles_for_existing_users) + ]
1bb1ececfcd548d52a28b713f4ee7eb4e710da85
keras_tf_multigpu/examples/fchollet_inception3_multigpu.py
keras_tf_multigpu/examples/fchollet_inception3_multigpu.py
import tensorflow as tf from keras.applications import InceptionV3 from keras.utils import multi_gpu_model import numpy as np num_samples = 1000 height = 224 width = 224 num_classes = 1000 gpu_count = 2 # Instantiate the base model # (here, we do it on CPU, which is optional). with tf.device('/cpu:0' if gpu_count > 1 else '/gpu:0'): model = InceptionV3(weights=None, input_shape=(height, width, 3), classes=num_classes) # Replicates the model on N GPUs. # This assumes that your machine has N available GPUs. if gpu_count > 1: parallel_model = multi_gpu_model(model, gpus=gpu_count) else: parallel_model = model parallel_model.compile(loss='categorical_crossentropy', optimizer='rmsprop') # Generate dummy data. x = np.random.random((num_samples, height, width, 3)) y = np.random.random((num_samples, num_classes)) # This `fit` call will be distributed on N GPUs. # Since the batch size is N*32, each GPU will process 32 samples. parallel_model.fit(x, y, epochs=20, batch_size=32 * gpu_count)
Add an example of using fchollet multi_gpu_model on InceptionV3.
Add an example of using fchollet multi_gpu_model on InceptionV3. Adapted from Keras example - parameterized number of GPUs, fixed imports, different model class, fixed device placement for single GPU.
Python
mit
rossumai/keras-multi-gpu,rossumai/keras-multi-gpu
--- +++ @@ -0,0 +1,35 @@ +import tensorflow as tf +from keras.applications import InceptionV3 +from keras.utils import multi_gpu_model +import numpy as np + +num_samples = 1000 +height = 224 +width = 224 +num_classes = 1000 + +gpu_count = 2 + +# Instantiate the base model +# (here, we do it on CPU, which is optional). +with tf.device('/cpu:0' if gpu_count > 1 else '/gpu:0'): + model = InceptionV3(weights=None, + input_shape=(height, width, 3), + classes=num_classes) + +# Replicates the model on N GPUs. +# This assumes that your machine has N available GPUs. +if gpu_count > 1: + parallel_model = multi_gpu_model(model, gpus=gpu_count) +else: + parallel_model = model +parallel_model.compile(loss='categorical_crossentropy', + optimizer='rmsprop') + +# Generate dummy data. +x = np.random.random((num_samples, height, width, 3)) +y = np.random.random((num_samples, num_classes)) + +# This `fit` call will be distributed on N GPUs. +# Since the batch size is N*32, each GPU will process 32 samples. +parallel_model.fit(x, y, epochs=20, batch_size=32 * gpu_count)
8c7fc2382db0ec9c901f6c2c2b00971f3ee7c3cc
logintokens/tests/test_backends.py
logintokens/tests/test_backends.py
"""logintokens app unittests for backends """ from time import sleep from django.test import TestCase, Client from django.contrib.auth import get_user_model, authenticate from logintokens.tokens import default_token_generator USER = get_user_model() class EmailOnlyAuthenticationBackendTest(TestCase): """Tests for email only authentication backend """ def setUp(self): self.client = Client() self.generator = default_token_generator self.new_username = 'newvisitor' self.existing_user = USER._default_manager.create_user('existinguser') def test_different_tokens_usable(self): """Two differing tokens should both be usabe to authenticate. """ username = self.existing_user.get_username() token1 = self.generator.make_token(username) sleep(1) token2 = self.generator.make_token(username) self.assertNotEqual(token1, token2) self.assertEqual(authenticate(token=token1), self.existing_user) self.assertEqual(authenticate(token=token2), self.existing_user) def test_login_invalidates_tokens(self): """Tokens generated before a successful login should become invalid. """ username = self.existing_user.get_username() token1 = self.generator.make_token(username) sleep(1) token2 = self.generator.make_token(username) self.assertNotEqual(token1, token2) self.client.force_login(self.existing_user) self.assertIsNone(authenticate(token=token1)) self.assertIsNone(authenticate(token=token2)) def test_new_visitor_creates_user(self): """Using a token from a new visitor should create their user object. """ token = self.generator.make_token(self.new_username) user = authenticate(token=token) self.assertIsInstance(user, USER)
Add tests for custom authentication backend
Add tests for custom authentication backend
Python
mit
randomic/aniauth-tdd,randomic/aniauth-tdd
--- +++ @@ -0,0 +1,60 @@ +"""logintokens app unittests for backends + +""" +from time import sleep + +from django.test import TestCase, Client +from django.contrib.auth import get_user_model, authenticate + +from logintokens.tokens import default_token_generator + + +USER = get_user_model() + + +class EmailOnlyAuthenticationBackendTest(TestCase): + """Tests for email only authentication backend + + """ + def setUp(self): + self.client = Client() + self.generator = default_token_generator + self.new_username = 'newvisitor' + self.existing_user = USER._default_manager.create_user('existinguser') + + def test_different_tokens_usable(self): + """Two differing tokens should both be usabe to authenticate. + + """ + username = self.existing_user.get_username() + token1 = self.generator.make_token(username) + sleep(1) + token2 = self.generator.make_token(username) + + self.assertNotEqual(token1, token2) + self.assertEqual(authenticate(token=token1), self.existing_user) + self.assertEqual(authenticate(token=token2), self.existing_user) + + def test_login_invalidates_tokens(self): + """Tokens generated before a successful login should become invalid. + + """ + username = self.existing_user.get_username() + token1 = self.generator.make_token(username) + sleep(1) + token2 = self.generator.make_token(username) + + self.assertNotEqual(token1, token2) + + self.client.force_login(self.existing_user) + + self.assertIsNone(authenticate(token=token1)) + self.assertIsNone(authenticate(token=token2)) + + def test_new_visitor_creates_user(self): + """Using a token from a new visitor should create their user object. + + """ + token = self.generator.make_token(self.new_username) + user = authenticate(token=token) + self.assertIsInstance(user, USER)
54ca48a2b8cbd53cd6506fdbce47d16f03a28a7d
tests/test_sorting_and_searching/test_bubble_sort.py
tests/test_sorting_and_searching/test_bubble_sort.py
import unittest from aids.sorting_and_searching.bubble_sort import bubble_sort class BubbleSortTestCase(unittest.TestCase): ''' Unit tests for bubble sort ''' def setUp(self): self.example_1 = [2, 5, 4, 3, 1] def test_bubble_sort(self): bubble_sort(self.example_1) self.assertEqual(self.example_1,[1,2,3,4,5]) def tearDown(self): pass if __name__ == '__main__': unittest.main()
Add unit tests for bubble sort
Add unit tests for bubble sort
Python
mit
ueg1990/aids
--- +++ @@ -0,0 +1,23 @@ +import unittest + +from aids.sorting_and_searching.bubble_sort import bubble_sort + +class BubbleSortTestCase(unittest.TestCase): + ''' + Unit tests for bubble sort + + ''' + + def setUp(self): + self.example_1 = [2, 5, 4, 3, 1] + + def test_bubble_sort(self): + bubble_sort(self.example_1) + self.assertEqual(self.example_1,[1,2,3,4,5]) + + + def tearDown(self): + pass + +if __name__ == '__main__': + unittest.main()
67596d081059a004e5f7ab15f7972773fdf2f15e
tests/syft/grid/messages/setup_msg_test.py
tests/syft/grid/messages/setup_msg_test.py
# syft absolute import syft as sy from syft.core.io.address import Address from syft.grid.messages.setup_messages import CreateInitialSetUpMessage from syft.grid.messages.setup_messages import CreateInitialSetUpResponse from syft.grid.messages.setup_messages import GetSetUpMessage from syft.grid.messages.setup_messages import GetSetUpResponse def test_create_initial_setup_message_serde() -> None: bob_vm = sy.VirtualMachine(name="Bob") target = Address(name="Alice") request_content = { "settings": { "cloud-admin-token" : "d84we35ad3a1d59a84sd9", "cloud-credentials": "<cloud-credentials.pem>", "infra": { "autoscaling": True, "triggers": { "memory": "50", "vCPU": "80" } }, } } msg = CreateInitialSetUpMessage( address=target, content= request_content, reply_to=bob_vm.address, ) blob = msg.serialize() msg2 = sy.deserialize(blob=blob) assert msg.id == msg2.id assert msg.address == target assert msg.content == msg2.content assert msg == msg2 def test_create_initial_setup_response_serde() -> None: target = Address(name="Alice") request_content = {"msg": "Initial setup registered successfully!"} msg = CreateInitialSetUpResponse( address=target, success=True, content= request_content, ) blob = msg.serialize() msg2 = sy.deserialize(blob=blob) assert msg.id == msg2.id assert msg.address == target assert msg.content == msg2.content assert msg == msg2 def test_get_initial_setup_message_serde() -> None: bob_vm = sy.VirtualMachine(name="Bob") target = Address(name="Alice") request_content = {} msg = GetSetUpMessage( address=target, content=request_content, reply_to=bob_vm.address, ) blob = msg.serialize() msg2 = sy.deserialize(blob=blob) assert msg.id == msg2.id assert msg.address == target assert msg.content == msg2.content assert msg == msg2 def test_delete_worker_response_serde() -> None: target = Address(name="Alice") content = { "settings": { "cloud-admin-token" : "d84we35ad3a1d59a84sd9", "cloud-credentials": "<cloud-credentials.pem>", "infra": { "autoscaling": True, "triggers": { "memory": "50", "vCPU": "80" } }, } } msg = GetSetUpResponse( success=True, address=target, content=content, ) blob = msg.serialize() msg2 = sy.deserialize(blob=blob) assert msg.id == msg2.id assert msg.address == target assert msg.content == msg2.content assert msg == msg2
ADD PyGrid SetupService message tests
ADD PyGrid SetupService message tests
Python
apache-2.0
OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft
--- +++ @@ -0,0 +1,106 @@ +# syft absolute +import syft as sy +from syft.core.io.address import Address +from syft.grid.messages.setup_messages import CreateInitialSetUpMessage +from syft.grid.messages.setup_messages import CreateInitialSetUpResponse +from syft.grid.messages.setup_messages import GetSetUpMessage +from syft.grid.messages.setup_messages import GetSetUpResponse + + +def test_create_initial_setup_message_serde() -> None: + bob_vm = sy.VirtualMachine(name="Bob") + target = Address(name="Alice") + + request_content = { + "settings": { + "cloud-admin-token" : "d84we35ad3a1d59a84sd9", + "cloud-credentials": "<cloud-credentials.pem>", + "infra": { + "autoscaling": True, + "triggers": { + "memory": "50", + "vCPU": "80" + } + }, + } + } + msg = CreateInitialSetUpMessage( + address=target, + content= request_content, + reply_to=bob_vm.address, + ) + + blob = msg.serialize() + msg2 = sy.deserialize(blob=blob) + + assert msg.id == msg2.id + assert msg.address == target + assert msg.content == msg2.content + assert msg == msg2 + +def test_create_initial_setup_response_serde() -> None: + target = Address(name="Alice") + + request_content = {"msg": "Initial setup registered successfully!"} + msg = CreateInitialSetUpResponse( + address=target, + success=True, + content= request_content, + ) + + blob = msg.serialize() + msg2 = sy.deserialize(blob=blob) + + assert msg.id == msg2.id + assert msg.address == target + assert msg.content == msg2.content + assert msg == msg2 + +def test_get_initial_setup_message_serde() -> None: + bob_vm = sy.VirtualMachine(name="Bob") + target = Address(name="Alice") + + request_content = {} + msg = GetSetUpMessage( + address=target, + content=request_content, + reply_to=bob_vm.address, + ) + + blob = msg.serialize() + msg2 = sy.deserialize(blob=blob) + + assert msg.id == msg2.id + assert msg.address == target + assert msg.content == msg2.content + assert msg == msg2 + +def test_delete_worker_response_serde() -> None: + target = Address(name="Alice") + + content = { + "settings": { + "cloud-admin-token" : "d84we35ad3a1d59a84sd9", + "cloud-credentials": "<cloud-credentials.pem>", + "infra": { + "autoscaling": True, + "triggers": { + "memory": "50", + "vCPU": "80" + } + }, + } + } + msg = GetSetUpResponse( + success=True, + address=target, + content=content, + ) + + blob = msg.serialize() + msg2 = sy.deserialize(blob=blob) + + assert msg.id == msg2.id + assert msg.address == target + assert msg.content == msg2.content + assert msg == msg2
478072e8350d03655364ea9147bbe21bafabbcce
geotrek/feedback/tests/test_template_tags.py
geotrek/feedback/tests/test_template_tags.py
from datetime import datetime from django.test import TestCase from geotrek.authent.tests.factories import UserFactory, UserProfileFactory from geotrek.feedback.templatetags.feedback_tags import ( predefined_emails, resolved_intervention_info, status_ids_and_colors) from geotrek.feedback.tests.factories import (PredefinedEmailFactory, ReportStatusFactory) from geotrek.maintenance.tests.factories import ReportInterventionFactory class TestFeedbackTemplateTags(TestCase): @classmethod def setUpTestData(cls): cls.user1 = UserFactory(username="CCCC") UserProfileFactory.create(user=cls.user1, extended_username="Communauté des Communes des Communautés Communataires") cls.user2 = UserFactory(username="Kurt") UserProfileFactory.create(user=cls.user2) solved_status = ReportStatusFactory(identifier='solved_intervention', color="#448654") cls.intervention_solved_1 = ReportInterventionFactory(date=datetime(year=1997, month=4, day=4).date()) cls.report_1 = cls.intervention_solved_1.target cls.report_1.status = solved_status cls.report_1.assigned_user = cls.user1 cls.report_1.save() cls.intervention_solved_2 = ReportInterventionFactory(date=datetime(year=1997, month=5, day=4).date()) cls.report_2 = cls.intervention_solved_2.target cls.report_2.status = solved_status cls.report_2.assigned_user = cls.user2 cls.report_2.save() cls.email1 = PredefinedEmailFactory() cls.email2 = PredefinedEmailFactory() def test_resolved_intervention_username(self): self.assertEqual( "{\"date\": \"04/04/1997\", \"username\": \"Communaut\\u00e9 des Communes des Communaut\\u00e9s Communataires\"}", resolved_intervention_info(self.report_1) ) self.assertEqual( "{\"date\": \"04/05/1997\", \"username\": \"Kurt\"}", resolved_intervention_info(self.report_2) ) def test_status_ids_and_colors(self): self.assertEqual( "{\"1\": {\"id\": \"solved_intervention\", \"color\": \"#448654\"}, \"2\": {\"id\": \"ID 1\", \"color\": \"#444444\"}, \"3\": {\"id\": \"ID 2\", \"color\": \"#444444\"}}", status_ids_and_colors() ) def test_predefined_emails(self): self.assertEqual( "{\"1\": {\"label\": \"Predefined Email 0\", \"text\": \"Some email body content 0\"}, \"2\": {\"label\": \"Predefined Email 1\", \"text\": \"Some email body content 1\"}}", predefined_emails() )
Add tests for template tags
Add tests for template tags
Python
bsd-2-clause
makinacorpus/Geotrek,makinacorpus/Geotrek,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek
--- +++ @@ -0,0 +1,55 @@ +from datetime import datetime + +from django.test import TestCase + +from geotrek.authent.tests.factories import UserFactory, UserProfileFactory +from geotrek.feedback.templatetags.feedback_tags import ( + predefined_emails, resolved_intervention_info, status_ids_and_colors) +from geotrek.feedback.tests.factories import (PredefinedEmailFactory, + ReportStatusFactory) +from geotrek.maintenance.tests.factories import ReportInterventionFactory + + +class TestFeedbackTemplateTags(TestCase): + + @classmethod + def setUpTestData(cls): + cls.user1 = UserFactory(username="CCCC") + UserProfileFactory.create(user=cls.user1, extended_username="Communauté des Communes des Communautés Communataires") + cls.user2 = UserFactory(username="Kurt") + UserProfileFactory.create(user=cls.user2) + solved_status = ReportStatusFactory(identifier='solved_intervention', color="#448654") + cls.intervention_solved_1 = ReportInterventionFactory(date=datetime(year=1997, month=4, day=4).date()) + cls.report_1 = cls.intervention_solved_1.target + cls.report_1.status = solved_status + cls.report_1.assigned_user = cls.user1 + cls.report_1.save() + cls.intervention_solved_2 = ReportInterventionFactory(date=datetime(year=1997, month=5, day=4).date()) + cls.report_2 = cls.intervention_solved_2.target + cls.report_2.status = solved_status + cls.report_2.assigned_user = cls.user2 + cls.report_2.save() + cls.email1 = PredefinedEmailFactory() + cls.email2 = PredefinedEmailFactory() + + def test_resolved_intervention_username(self): + self.assertEqual( + "{\"date\": \"04/04/1997\", \"username\": \"Communaut\\u00e9 des Communes des Communaut\\u00e9s Communataires\"}", + resolved_intervention_info(self.report_1) + ) + self.assertEqual( + "{\"date\": \"04/05/1997\", \"username\": \"Kurt\"}", + resolved_intervention_info(self.report_2) + ) + + def test_status_ids_and_colors(self): + self.assertEqual( + "{\"1\": {\"id\": \"solved_intervention\", \"color\": \"#448654\"}, \"2\": {\"id\": \"ID 1\", \"color\": \"#444444\"}, \"3\": {\"id\": \"ID 2\", \"color\": \"#444444\"}}", + status_ids_and_colors() + ) + + def test_predefined_emails(self): + self.assertEqual( + "{\"1\": {\"label\": \"Predefined Email 0\", \"text\": \"Some email body content 0\"}, \"2\": {\"label\": \"Predefined Email 1\", \"text\": \"Some email body content 1\"}}", + predefined_emails() + )
8355cb358d14589a194926d37beeb5af7af2a591
falmer/events/migrations/0012_auto_20170905_1208.py
falmer/events/migrations/0012_auto_20170905_1208.py
# -*- coding: utf-8 -*- # Generated by Django 1.11.3 on 2017-09-05 11:08 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('events', '0011_auto_20170905_1028'), ] operations = [ migrations.AlterField( model_name='mslevent', name='image_url', field=models.URLField(max_length=2000), ), migrations.AlterField( model_name='mslevent', name='url', field=models.URLField(max_length=2000), ), ]
Increase event image url limit from 200
Increase event image url limit from 200
Python
mit
sussexstudent/falmer,sussexstudent/falmer,sussexstudent/falmer,sussexstudent/falmer
--- +++ @@ -0,0 +1,25 @@ +# -*- coding: utf-8 -*- +# Generated by Django 1.11.3 on 2017-09-05 11:08 +from __future__ import unicode_literals + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('events', '0011_auto_20170905_1028'), + ] + + operations = [ + migrations.AlterField( + model_name='mslevent', + name='image_url', + field=models.URLField(max_length=2000), + ), + migrations.AlterField( + model_name='mslevent', + name='url', + field=models.URLField(max_length=2000), + ), + ]
a16b4401f37f08d8cb5e1f9ec1b7d4a3221360ab
test/test_regular_extrusion.py
test/test_regular_extrusion.py
# -*- coding: utf-8 -*- """Creates regular cube mesh by extrusion. """ import pygmsh from helpers import compute_volume def test(): x = 5 y = 4 z = 3 x_layers = 10 y_layers = 5 z_layers = 3 geom = pygmsh.built_in.Geometry() p = geom.add_point([0, 0, 0], 1) _, l, _ = geom.extrude(p, [x, 0, 0], num_layers=x_layers) _, s, _ = geom.extrude(l, [0, y, 0], num_layers=y_layers) geom.extrude(s, [0, 0, z], num_layers=z_layers) points, cells, _, _, _ = pygmsh.generate_mesh(geom) ref_vol = x * y * z assert abs(compute_volume(points, cells) - ref_vol) < 1.0e-2 * ref_vol # Each grid-cell from layered extrusion will result in 6 tetrahedrons. ref_tetras = 6 * x_layers * y_layers * z_layers assert len(cells["tetra"]) == ref_tetras return points, cells if __name__ == "__main__": import meshio meshio.write_points_cells("cube.vtu", *test())
Add test case for regular extrusion
Add test case for regular extrusion
Python
bsd-3-clause
nschloe/python4gmsh
--- +++ @@ -0,0 +1,36 @@ +# -*- coding: utf-8 -*- +"""Creates regular cube mesh by extrusion. +""" +import pygmsh + +from helpers import compute_volume + + +def test(): + x = 5 + y = 4 + z = 3 + x_layers = 10 + y_layers = 5 + z_layers = 3 + geom = pygmsh.built_in.Geometry() + p = geom.add_point([0, 0, 0], 1) + _, l, _ = geom.extrude(p, [x, 0, 0], num_layers=x_layers) + _, s, _ = geom.extrude(l, [0, y, 0], num_layers=y_layers) + geom.extrude(s, [0, 0, z], num_layers=z_layers) + points, cells, _, _, _ = pygmsh.generate_mesh(geom) + + ref_vol = x * y * z + assert abs(compute_volume(points, cells) - ref_vol) < 1.0e-2 * ref_vol + + # Each grid-cell from layered extrusion will result in 6 tetrahedrons. + ref_tetras = 6 * x_layers * y_layers * z_layers + assert len(cells["tetra"]) == ref_tetras + + return points, cells + + +if __name__ == "__main__": + import meshio + + meshio.write_points_cells("cube.vtu", *test())
7bd6f3e7751deecfc3cd555fc071d722c856802c
chips/compiler/builtins.py
chips/compiler/builtins.py
#!/usr/bin/env python """Support Library for builtin Functionality""" __author__ = "Jon Dawson" __copyright__ = "Copyright (C) 2013, Jonathan P Dawson" __version__ = "0.1" builtins=""" unsigned unsigned_divide_xxxx(unsigned dividend, unsigned divisor){ unsigned denom = divisor; unsigned bit = 1; unsigned quotient = 0; if( denom > dividend ) return 0; if( denom == dividend ) return 1; while(denom <= dividend){ denom <<= 1; bit <<= 1; } denom >>= 1; bit >>= 1; while(bit){ if(dividend >= denom){ dividend -= denom; quotient |= bit; } bit >>= 1; denom >>= 1; } return quotient; } int divide_xxxx(int dividend, int divisor){ unsigned udividend, udivisor, uquotient; unsigned dividend_sign, divisor_sign, quotient_sign; dividend_sign = dividend & 0x8000u; divisor_sign = divisor & 0x8000u; quotient_sign = dividend_sign ^ divisor_sign; udividend = dividend_sign ? -dividend : dividend; udivisor = divisor_sign ? -divisor : divisor; uquotient = unsigned_divide_xxxx(udividend, udivisor); return quotient_sign ? -uquotient : uquotient; } long unsigned long_unsigned_divide_xxxx(long unsigned dividend, long unsigned divisor){ long unsigned denom = divisor; long unsigned bit = 1; long unsigned quotient = 0; if( denom > dividend ) return 0; if( denom == dividend ) return 1; while(denom <= dividend){ denom <<= 1; bit <<= 1; } denom >>= 1; bit >>= 1; while(bit){ if(dividend >= denom){ dividend -= denom; quotient |= bit; } bit >>= 1; denom >>= 1; } return quotient; } long int long_divide_xxxx(long int dividend, long int divisor){ long unsigned udividend, udivisor, uquotient; long unsigned dividend_sign, divisor_sign, quotient_sign; dividend_sign = dividend & 0x80000000ul; divisor_sign = divisor & 0x80000000ul; quotient_sign = dividend_sign ^ divisor_sign; udividend = dividend_sign ? -dividend : dividend; udivisor = divisor_sign ? -divisor : divisor; uquotient = long_unsigned_divide_xxxx(udividend, udivisor); return quotient_sign ? -uquotient : uquotient; } """
Implement division using built in library function
Implement division using built in library function
Python
mit
dawsonjon/Chips-2.0,dawsonjon/Chips-2.0,dawsonjon/Chips-2.0,dawsonjon/Chips-2.0,dawsonjon/Chips-2.0
--- +++ @@ -0,0 +1,80 @@ +#!/usr/bin/env python +"""Support Library for builtin Functionality""" + +__author__ = "Jon Dawson" +__copyright__ = "Copyright (C) 2013, Jonathan P Dawson" +__version__ = "0.1" + +builtins=""" + +unsigned unsigned_divide_xxxx(unsigned dividend, unsigned divisor){ + unsigned denom = divisor; + unsigned bit = 1; + unsigned quotient = 0; + if( denom > dividend ) return 0; + if( denom == dividend ) return 1; + while(denom <= dividend){ + denom <<= 1; + bit <<= 1; + } + denom >>= 1; + bit >>= 1; + while(bit){ + if(dividend >= denom){ + dividend -= denom; + quotient |= bit; + } + bit >>= 1; + denom >>= 1; + } + return quotient; +} + +int divide_xxxx(int dividend, int divisor){ + unsigned udividend, udivisor, uquotient; + unsigned dividend_sign, divisor_sign, quotient_sign; + dividend_sign = dividend & 0x8000u; + divisor_sign = divisor & 0x8000u; + quotient_sign = dividend_sign ^ divisor_sign; + udividend = dividend_sign ? -dividend : dividend; + udivisor = divisor_sign ? -divisor : divisor; + uquotient = unsigned_divide_xxxx(udividend, udivisor); + return quotient_sign ? -uquotient : uquotient; +} + +long unsigned long_unsigned_divide_xxxx(long unsigned dividend, long unsigned divisor){ + long unsigned denom = divisor; + long unsigned bit = 1; + long unsigned quotient = 0; + if( denom > dividend ) return 0; + if( denom == dividend ) return 1; + while(denom <= dividend){ + denom <<= 1; + bit <<= 1; + } + denom >>= 1; + bit >>= 1; + while(bit){ + if(dividend >= denom){ + dividend -= denom; + quotient |= bit; + } + bit >>= 1; + denom >>= 1; + } + return quotient; +} + +long int long_divide_xxxx(long int dividend, long int divisor){ + long unsigned udividend, udivisor, uquotient; + long unsigned dividend_sign, divisor_sign, quotient_sign; + dividend_sign = dividend & 0x80000000ul; + divisor_sign = divisor & 0x80000000ul; + quotient_sign = dividend_sign ^ divisor_sign; + udividend = dividend_sign ? -dividend : dividend; + udivisor = divisor_sign ? -divisor : divisor; + uquotient = long_unsigned_divide_xxxx(udividend, udivisor); + return quotient_sign ? -uquotient : uquotient; +} + +"""
9316bc07c77e2f51332a40bf430cef117f4d89e1
util/check_dockerfile_coverage.py
util/check_dockerfile_coverage.py
import yaml import os import pathlib2 import itertools import argparse import logging import sys TRAVIS_BUILD_DIR = os.environ.get("TRAVIS_BUILD_DIR") CONFIG_FILE_PATH = pathlib2.Path(TRAVIS_BUILD_DIR, "util", "parsefiles_config.yml") LOGGER = logging.getLogger(__name__) def check_coverage(containers): # open config file containing container weights config_file_path = pathlib2.Path(CONFIG_FILE_PATH) with (config_file_path.open(mode='r')) as file: try: config = yaml.load(file) except yaml.YAMLError, exc: LOGGER.error("error in configuration file: %s" % str(exc)) sys.exit(1) # get container weights weights = config.get("weights") # convert all containers in config file to a list of tuples (<container>, <weight>) weights_list = [x.items() for x in weights] weights_list = list(itertools.chain.from_iterable(weights_list)) # performs intersection between weighted containers and input containers used_containers = [x for x in weights_list if x[0] in containers] # determine which Dockerfiles are not covered; i.e. the set difference of the Dockerfiles to build minus the Dockerfile # available to be built is non-empty uncovered = set(containers) - set([x[0] for x in used_containers]) # exit with error code if uncovered Dockerfiles exist if uncovered: LOGGER.error("The following Dockerfiles are not described in the parsefiles_config.yml file: {}. Please see the following documentation on how to add Dockerfile ranks to the configuration file: {}".format(uncovered, "https://github.com/edx/configuration/blob/master/util/README.md")) sys.exit(1) def arg_parse(): parser = argparse.ArgumentParser(description = 'Given a list of containers as input and a number of shards, ' 'finds an approximation of the optimal distribution of the containers over the shards, provided a set of hard-coded weights ' 'in parsefiles_config.yml.') parser.add_argument('containers', help = "the Dockerfiles that need to be built as the result of some commit change and whose coverage is checked") return parser.parse_args() if __name__ == '__main__': args = arg_parse() # configure logging logging.basicConfig() containers = [] for word in args.containers.split(): containers.append(word) check_coverage(containers)
Add script to check for Dockerfile coverage
Add script to check for Dockerfile coverage
Python
agpl-3.0
rue89-tech/configuration,stvstnfrd/configuration,michaelsteiner19/open-edx-configuration,armaan/edx-configuration,hastexo/edx-configuration,armaan/edx-configuration,stvstnfrd/configuration,gsehub/configuration,stvstnfrd/configuration,hks-epod/configuration,open-craft/configuration,edx/configuration,open-craft/configuration,armaan/edx-configuration,proversity-org/configuration,stvstnfrd/configuration,hks-epod/configuration,michaelsteiner19/open-edx-configuration,edx/configuration,gsehub/configuration,edx/configuration,Stanford-Online/configuration,hastexo/edx-configuration,gsehub/configuration,hastexo/edx-configuration,open-craft/configuration,appsembler/configuration,rue89-tech/configuration,arbrandes/edx-configuration,michaelsteiner19/open-edx-configuration,open-craft/configuration,EDUlib/configuration,nunpa/configuration,nunpa/configuration,arbrandes/edx-configuration,mitodl/configuration,mitodl/configuration,Stanford-Online/configuration,rue89-tech/configuration,hks-epod/configuration,Stanford-Online/configuration,jorgeomarmh/configuration,appsembler/configuration,jorgeomarmh/configuration,EDUlib/configuration,armaan/edx-configuration,EDUlib/configuration,nunpa/configuration,stvstnfrd/configuration,hastexo/edx-configuration,gsehub/configuration,proversity-org/configuration,michaelsteiner19/open-edx-configuration,proversity-org/configuration,arbrandes/edx-configuration,arbrandes/edx-configuration,Stanford-Online/configuration,edx/configuration,hks-epod/configuration,jorgeomarmh/configuration,proversity-org/configuration,appsembler/configuration,appsembler/configuration,nunpa/configuration,mitodl/configuration,EDUlib/configuration,mitodl/configuration,jorgeomarmh/configuration,rue89-tech/configuration,rue89-tech/configuration,EDUlib/configuration,hks-epod/configuration,Stanford-Online/configuration,proversity-org/configuration,gsehub/configuration,hastexo/edx-configuration,arbrandes/edx-configuration
--- +++ @@ -0,0 +1,64 @@ +import yaml +import os +import pathlib2 +import itertools +import argparse +import logging +import sys + +TRAVIS_BUILD_DIR = os.environ.get("TRAVIS_BUILD_DIR") +CONFIG_FILE_PATH = pathlib2.Path(TRAVIS_BUILD_DIR, "util", "parsefiles_config.yml") +LOGGER = logging.getLogger(__name__) + +def check_coverage(containers): + # open config file containing container weights + config_file_path = pathlib2.Path(CONFIG_FILE_PATH) + + with (config_file_path.open(mode='r')) as file: + try: + config = yaml.load(file) + except yaml.YAMLError, exc: + LOGGER.error("error in configuration file: %s" % str(exc)) + sys.exit(1) + + # get container weights + weights = config.get("weights") + + # convert all containers in config file to a list of tuples (<container>, <weight>) + weights_list = [x.items() for x in weights] + weights_list = list(itertools.chain.from_iterable(weights_list)) + + # performs intersection between weighted containers and input containers + used_containers = [x for x in weights_list if x[0] in containers] + + # determine which Dockerfiles are not covered; i.e. the set difference of the Dockerfiles to build minus the Dockerfile + # available to be built is non-empty + uncovered = set(containers) - set([x[0] for x in used_containers]) + + # exit with error code if uncovered Dockerfiles exist + if uncovered: + LOGGER.error("The following Dockerfiles are not described in the parsefiles_config.yml file: {}. Please see the following documentation on how to add Dockerfile ranks to the configuration file: {}".format(uncovered, "https://github.com/edx/configuration/blob/master/util/README.md")) + sys.exit(1) + +def arg_parse(): + + parser = argparse.ArgumentParser(description = 'Given a list of containers as input and a number of shards, ' + 'finds an approximation of the optimal distribution of the containers over the shards, provided a set of hard-coded weights ' + 'in parsefiles_config.yml.') + parser.add_argument('containers', help = "the Dockerfiles that need to be built as the result of some commit change and whose coverage is checked") + + return parser.parse_args() + +if __name__ == '__main__': + + args = arg_parse() + + # configure logging + logging.basicConfig() + + containers = [] + + for word in args.containers.split(): + containers.append(word) + + check_coverage(containers)
805708048f493ca538a9e0b8d9d40ae1d4baf2c3
keepalive-race/keep-alive-race.py
keepalive-race/keep-alive-race.py
#!/usr/bin/python3 """ This script demonstrates a race condition with HTTP/1.1 keepalive """ import decimal import json import subprocess import time import threading import requests requests.packages.urllib3.disable_warnings() CREDS = json.loads(subprocess.check_output( "openstack --os-cloud devstack token issue -f json".split(), ).decode()) URL = 'https://10.0.1.44:8774/v2/%s/servers/detail' % (CREDS['project_id']) def decimal_range(x, y, jump): x = decimal.Decimal(x) y = decimal.Decimal(y) jump = decimal.Decimal(jump) while x < y: yield float(x) x += jump def get(exit): for delay in decimal_range(4.95, 4.96, 0.005): session = requests.Session() if exit.is_set(): return for i in range(10): if exit.is_set(): return time.sleep(delay) headers = { 'User-Agent': 'timeout-race/%s' % i, 'X-Auth-Token': CREDS['id'] } try: session.get(URL, verify=False, headers=headers) except Exception as e: print(e) exit.set() threads = [] exit = threading.Event() for i in range(50): threads.append(threading.Thread(target=get,args=(exit,))) for thread in threads: thread.start() for thread in threads: thread.join()
Add a tool to reproduce HTTP KeepAlive races in OpenStack gate jobs.
Add a tool to reproduce HTTP KeepAlive races in OpenStack gate jobs. Based on https://github.com/mikem23/keepalive-race Change-Id: I11f66bf39c6cc2609ee0dbff97ac7b104767ac2b
Python
apache-2.0
JordanP/openstack-snippets,JordanP/openstack-snippets
--- +++ @@ -0,0 +1,62 @@ +#!/usr/bin/python3 +""" +This script demonstrates a race condition with HTTP/1.1 keepalive +""" +import decimal +import json +import subprocess +import time +import threading + +import requests +requests.packages.urllib3.disable_warnings() + +CREDS = json.loads(subprocess.check_output( + "openstack --os-cloud devstack token issue -f json".split(), +).decode()) +URL = 'https://10.0.1.44:8774/v2/%s/servers/detail' % (CREDS['project_id']) + + +def decimal_range(x, y, jump): + x = decimal.Decimal(x) + y = decimal.Decimal(y) + jump = decimal.Decimal(jump) + while x < y: + yield float(x) + x += jump + + +def get(exit): + for delay in decimal_range(4.95, 4.96, 0.005): + session = requests.Session() + + if exit.is_set(): + return + + for i in range(10): + + if exit.is_set(): + return + + time.sleep(delay) + headers = { + 'User-Agent': 'timeout-race/%s' % i, + 'X-Auth-Token': CREDS['id'] + } + try: + session.get(URL, verify=False, headers=headers) + except Exception as e: + print(e) + exit.set() + + +threads = [] +exit = threading.Event() +for i in range(50): + threads.append(threading.Thread(target=get,args=(exit,))) + +for thread in threads: + thread.start() + +for thread in threads: + thread.join()
48da7ceb86387d3cb6fd53f50110232813123ecc
tests/pytests/unit/roster/test_ansible.py
tests/pytests/unit/roster/test_ansible.py
import pytest import salt.roster.ansible as ansible from tests.support.mock import patch @pytest.mark.xfail @pytest.mark.parametrize( "which_value", [False, None], ) def test_virtual_returns_False_if_ansible_inventory_doesnt_exist(which_value): with patch("salt.utils.path.which", autospec=True, return_value=which_value): assert ansible.__virtual__() == (False, "Install `ansible` to use inventory")
Add tests for ansible roster virtual
Add tests for ansible roster virtual
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
--- +++ @@ -0,0 +1,13 @@ +import pytest +import salt.roster.ansible as ansible +from tests.support.mock import patch + + +@pytest.mark.xfail +@pytest.mark.parametrize( + "which_value", + [False, None], +) +def test_virtual_returns_False_if_ansible_inventory_doesnt_exist(which_value): + with patch("salt.utils.path.which", autospec=True, return_value=which_value): + assert ansible.__virtual__() == (False, "Install `ansible` to use inventory")
59e546ae5afe22aab967e5376c8799e29ccbd86a
directoryFileContentCmp.py
directoryFileContentCmp.py
#! /usr/env/python import os import hashlib import sys bufsize = 65536 # Path1 = '/Users/kirkchambers/Desktop' # Path2 = '/Users/kirkchambers/DataSets' def generate_file_digests_for(path): path_set = set() for item in os.walk(path): (directory, _subdirectories, files) = item for file in files: if (file[0] == '.'): continue else: fqFilename = os.path.join(directory, file) path_set.add(generate_file_digest(fqFilename, file)) return path_set def generate_file_digest(fqFilename, shortFilename): hasher = hashlib.md5() with open(fqFilename, 'rb') as filestream: fileBuffer = filestream.read(bufsize) while len(fileBuffer) > 0: hasher.update(fileBuffer) fileBuffer = filestream.read(bufsize) # return "Filename:{file}\nHash:{hash}\nSize:{size}\n".format(file=fqFilename, hash=hasher.hexdigest(), size=os.path.getsize(fqFilename)) return (hasher.hexdigest(), fqFilename, os.path.getsize(fqFilename)) def usage(): print "file_list.py directory1 directory2" print "Prints out the files present in directory1 which are NOT present in directory2" if __name__ == "__main__": try: (_command, Path1, Path2) = sys.argv except: usage() exit(1) path_set_1 = generate_file_digests_for(Path1) path_set_2 = generate_file_digests_for(Path2) # union = path_set_1 | path_set_2 set_1_exclusives = path_set_1 - path_set_2 # set_2_exclusives = path_set_2 - path_set_1 # print "length of 1: {}".format(len(path_set_1)) # print "length of 2: {}".format(len(path_set_2)) # print "length of union: {}".format(len(union)) # print "length of set1 uniqs: {}".format(len(set_1_exclusives)) # print "length of set2 uniqs: {}".format(len(set_2_exclusives)) print "Files present in {path1} and not in {path2}:".format(path1=Path1, path2=Path2) for item in set_1_exclusives: print item[1]
Add the basic version of my file comparison script
Add the basic version of my file comparison script
Python
mit
kirkchambe/random_scripts
--- +++ @@ -0,0 +1,60 @@ +#! /usr/env/python +import os +import hashlib +import sys + +bufsize = 65536 +# Path1 = '/Users/kirkchambers/Desktop' +# Path2 = '/Users/kirkchambers/DataSets' + +def generate_file_digests_for(path): + path_set = set() + for item in os.walk(path): + (directory, _subdirectories, files) = item + for file in files: + if (file[0] == '.'): + continue + else: + fqFilename = os.path.join(directory, file) + path_set.add(generate_file_digest(fqFilename, file)) + return path_set + + +def generate_file_digest(fqFilename, shortFilename): + hasher = hashlib.md5() + with open(fqFilename, 'rb') as filestream: + fileBuffer = filestream.read(bufsize) + while len(fileBuffer) > 0: + hasher.update(fileBuffer) + fileBuffer = filestream.read(bufsize) + # return "Filename:{file}\nHash:{hash}\nSize:{size}\n".format(file=fqFilename, hash=hasher.hexdigest(), size=os.path.getsize(fqFilename)) + return (hasher.hexdigest(), fqFilename, os.path.getsize(fqFilename)) + + +def usage(): + print "file_list.py directory1 directory2" + print "Prints out the files present in directory1 which are NOT present in directory2" + + +if __name__ == "__main__": + try: + (_command, Path1, Path2) = sys.argv + except: + usage() + exit(1) + + path_set_1 = generate_file_digests_for(Path1) + path_set_2 = generate_file_digests_for(Path2) + + # union = path_set_1 | path_set_2 + set_1_exclusives = path_set_1 - path_set_2 + # set_2_exclusives = path_set_2 - path_set_1 + # print "length of 1: {}".format(len(path_set_1)) + # print "length of 2: {}".format(len(path_set_2)) + # print "length of union: {}".format(len(union)) + # print "length of set1 uniqs: {}".format(len(set_1_exclusives)) + # print "length of set2 uniqs: {}".format(len(set_2_exclusives)) + + print "Files present in {path1} and not in {path2}:".format(path1=Path1, path2=Path2) + for item in set_1_exclusives: + print item[1]
c3afc6c28530c3dfc3bd57d9a1841a60bf92ba4f
tools/perf/benchmarks/netsim_top25.py
tools/perf/benchmarks/netsim_top25.py
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from telemetry import test from perf_tools import page_cycler class NetsimTop25(test.Test): """Measures load time of the top 25 sites under simulated cable network.""" test = page_cycler.PageCycler test.clear_cache_before_each_run = True page_set = 'tools/perf/page_sets/top_25.json' options = { 'extra_wpr_args': [ '--shaping_type=proxy', '--net=cable' ], 'pageset_repeat': '5', }
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from telemetry import test from perf_tools import page_cycler class NetsimTop25(test.Test): """Measures load time of the top 25 sites under simulated cable network.""" test = page_cycler.PageCycler page_set = 'tools/perf/page_sets/top_25.json' options = { 'extra_wpr_args': [ '--shaping_type=proxy', '--net=cable' ], 'pageset_repeat': '5', } def __init__(self): super(NetsimTop25, self).__init__() self.test.clear_cache_before_each_run = True
Fix bug which caused page cyclers to always clear cache before load.
[Telemetry] Fix bug which caused page cyclers to always clear cache before load. Previously, the cache clearing bit would apply when the netsim benchmark was imported. This fixes it so that it only applies when it is used. BUG=256492 NOTRY=True TBR=dtu@chromium.org Review URL: https://codereview.chromium.org/18550003 git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@209708 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
mohamed--abdel-maksoud/chromium.src,patrickm/chromium.src,chuan9/chromium-crosswalk,jaruba/chromium.src,Chilledheart/chromium,littlstar/chromium.src,fujunwei/chromium-crosswalk,jaruba/chromium.src,chuan9/chromium-crosswalk,Pluto-tv/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk-efl,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,dushu1203/chromium.src,Jonekee/chromium.src,Just-D/chromium-1,ltilve/chromium,mohamed--abdel-maksoud/chromium.src,anirudhSK/chromium,Jonekee/chromium.src,Pluto-tv/chromium-crosswalk,dednal/chromium.src,M4sse/chromium.src,markYoungH/chromium.src,axinging/chromium-crosswalk,anirudhSK/chromium,ltilve/chromium,hgl888/chromium-crosswalk-efl,TheTypoMaster/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,markYoungH/chromium.src,chuan9/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Just-D/chromium-1,Just-D/chromium-1,Chilledheart/chromium,chuan9/chromium-crosswalk,Fireblend/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Just-D/chromium-1,patrickm/chromium.src,ChromiumWebApps/chromium,bright-sparks/chromium-spacewalk,ltilve/chromium,ondra-novak/chromium.src,Pluto-tv/chromium-crosswalk,dushu1203/chromium.src,dushu1203/chromium.src,TheTypoMaster/chromium-crosswalk,anirudhSK/chromium,littlstar/chromium.src,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,axinging/chromium-crosswalk,bright-sparks/chromium-spacewalk,ltilve/chromium,hgl888/chromium-crosswalk-efl,Chilledheart/chromium,M4sse/chromium.src,M4sse/chromium.src,bright-sparks/chromium-spacewalk,chuan9/chromium-crosswalk,patrickm/chromium.src,krieger-od/nwjs_chromium.src,mohamed--abdel-maksoud/chromium.src,hgl888/chromium-crosswalk,ondra-novak/chromium.src,littlstar/chromium.src,ondra-novak/chromium.src,krieger-od/nwjs_chromium.src,M4sse/chromium.src,Jonekee/chromium.src,dushu1203/chromium.src,chuan9/chromium-crosswalk,Chilledheart/chromium,axinging/chromium-crosswalk,jaruba/chromium.src,mohamed--abdel-maksoud/chromium.src,bright-sparks/chromium-spacewalk,dushu1203/chromium.src,patrickm/chromium.src,hgl888/chromium-crosswalk-efl,Jonekee/chromium.src,fujunwei/chromium-crosswalk,mogoweb/chromium-crosswalk,ondra-novak/chromium.src,anirudhSK/chromium,ondra-novak/chromium.src,Fireblend/chromium-crosswalk,anirudhSK/chromium,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,hgl888/chromium-crosswalk-efl,patrickm/chromium.src,axinging/chromium-crosswalk,ondra-novak/chromium.src,dednal/chromium.src,TheTypoMaster/chromium-crosswalk,ondra-novak/chromium.src,littlstar/chromium.src,bright-sparks/chromium-spacewalk,PeterWangIntel/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,dednal/chromium.src,markYoungH/chromium.src,Chilledheart/chromium,Pluto-tv/chromium-crosswalk,M4sse/chromium.src,mohamed--abdel-maksoud/chromium.src,crosswalk-project/chromium-crosswalk-efl,dushu1203/chromium.src,Just-D/chromium-1,ltilve/chromium,M4sse/chromium.src,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,anirudhSK/chromium,chuan9/chromium-crosswalk,Chilledheart/chromium,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,axinging/chromium-crosswalk,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,anirudhSK/chromium,bright-sparks/chromium-spacewalk,ChromiumWebApps/chromium,markYoungH/chromium.src,Pluto-tv/chromium-crosswalk,Jonekee/chromium.src,Jonekee/chromium.src,ltilve/chromium,ChromiumWebApps/chromium,hgl888/chromium-crosswalk,markYoungH/chromium.src,bright-sparks/chromium-spacewalk,mohamed--abdel-maksoud/chromium.src,littlstar/chromium.src,patrickm/chromium.src,M4sse/chromium.src,Jonekee/chromium.src,crosswalk-project/chromium-crosswalk-efl,ChromiumWebApps/chromium,Just-D/chromium-1,Jonekee/chromium.src,bright-sparks/chromium-spacewalk,dednal/chromium.src,PeterWangIntel/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,bright-sparks/chromium-spacewalk,littlstar/chromium.src,axinging/chromium-crosswalk,Just-D/chromium-1,ChromiumWebApps/chromium,mogoweb/chromium-crosswalk,ChromiumWebApps/chromium,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,ChromiumWebApps/chromium,dushu1203/chromium.src,Chilledheart/chromium,littlstar/chromium.src,dednal/chromium.src,littlstar/chromium.src,krieger-od/nwjs_chromium.src,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk,dednal/chromium.src,PeterWangIntel/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Chilledheart/chromium,ltilve/chromium,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk-efl,mogoweb/chromium-crosswalk,fujunwei/chromium-crosswalk,Jonekee/chromium.src,dushu1203/chromium.src,ChromiumWebApps/chromium,Fireblend/chromium-crosswalk,chuan9/chromium-crosswalk,anirudhSK/chromium,Jonekee/chromium.src,jaruba/chromium.src,M4sse/chromium.src,jaruba/chromium.src,markYoungH/chromium.src,mohamed--abdel-maksoud/chromium.src,crosswalk-project/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,ondra-novak/chromium.src,hgl888/chromium-crosswalk-efl,hgl888/chromium-crosswalk,jaruba/chromium.src,crosswalk-project/chromium-crosswalk-efl,Chilledheart/chromium,markYoungH/chromium.src,dednal/chromium.src,hgl888/chromium-crosswalk,Just-D/chromium-1,axinging/chromium-crosswalk,jaruba/chromium.src,ChromiumWebApps/chromium,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,Fireblend/chromium-crosswalk,M4sse/chromium.src,crosswalk-project/chromium-crosswalk-efl,anirudhSK/chromium,jaruba/chromium.src,dednal/chromium.src,dushu1203/chromium.src,dednal/chromium.src,fujunwei/chromium-crosswalk,anirudhSK/chromium,ltilve/chromium,PeterWangIntel/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,patrickm/chromium.src,Jonekee/chromium.src,krieger-od/nwjs_chromium.src,Fireblend/chromium-crosswalk,Pluto-tv/chromium-crosswalk,hgl888/chromium-crosswalk-efl,markYoungH/chromium.src,crosswalk-project/chromium-crosswalk-efl,axinging/chromium-crosswalk,hgl888/chromium-crosswalk-efl,krieger-od/nwjs_chromium.src,mohamed--abdel-maksoud/chromium.src,hgl888/chromium-crosswalk,chuan9/chromium-crosswalk,ChromiumWebApps/chromium,dednal/chromium.src,Fireblend/chromium-crosswalk,markYoungH/chromium.src,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,jaruba/chromium.src,anirudhSK/chromium,dushu1203/chromium.src,patrickm/chromium.src,mogoweb/chromium-crosswalk,jaruba/chromium.src,ltilve/chromium,PeterWangIntel/chromium-crosswalk,ChromiumWebApps/chromium,M4sse/chromium.src,axinging/chromium-crosswalk,markYoungH/chromium.src,mogoweb/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,markYoungH/chromium.src,fujunwei/chromium-crosswalk,fujunwei/chromium-crosswalk,M4sse/chromium.src,ondra-novak/chromium.src,anirudhSK/chromium,dednal/chromium.src,Just-D/chromium-1,patrickm/chromium.src,crosswalk-project/chromium-crosswalk-efl
--- +++ @@ -9,7 +9,6 @@ class NetsimTop25(test.Test): """Measures load time of the top 25 sites under simulated cable network.""" test = page_cycler.PageCycler - test.clear_cache_before_each_run = True page_set = 'tools/perf/page_sets/top_25.json' options = { 'extra_wpr_args': [ @@ -18,3 +17,7 @@ ], 'pageset_repeat': '5', } + + def __init__(self): + super(NetsimTop25, self).__init__() + self.test.clear_cache_before_each_run = True
fe145fd87db777d9eeb361688d502b1b3ec4b2e1
Transformation.py
Transformation.py
# -*- coding:utf-8 -*- # *************************************************************************** # Transformation.py # ------------------- # update : 2013-11-13 # copyright : (C) 2013 by Michaël Roy # email : microygh@gmail.com # *************************************************************************** # *************************************************************************** # * * # * This program is free software; you can redistribute it and/or modify * # * it under the terms of the GNU General Public License as published by * # * the Free Software Foundation; either version 2 of the License, or * # * (at your option) any later version. * # * * # *************************************************************************** # # External dependencies # from numpy import *
Add a new Model-View-Projection matrix tool.
Add a new Model-View-Projection matrix tool.
Python
mit
microy/MeshToolkit,microy/PyMeshToolkit,microy/MeshToolkit,microy/PyMeshToolkit
--- +++ @@ -0,0 +1,26 @@ +# -*- coding:utf-8 -*- + +# *************************************************************************** +# Transformation.py +# ------------------- +# update : 2013-11-13 +# copyright : (C) 2013 by Michaël Roy +# email : microygh@gmail.com +# *************************************************************************** + +# *************************************************************************** +# * * +# * This program is free software; you can redistribute it and/or modify * +# * it under the terms of the GNU General Public License as published by * +# * the Free Software Foundation; either version 2 of the License, or * +# * (at your option) any later version. * +# * * +# *************************************************************************** + + +# +# External dependencies +# +from numpy import * + +
68dbfedf90fb9e6c922971deaeccad148a258a70
tests/test_dynamic_ecore_extension.py
tests/test_dynamic_ecore_extension.py
import pytest from pyecore.ecore import * import pyecore.ecore as ecore from ordered_set import OrderedSet def test__EModelElement_extension(): A = EClass('A', superclass=(EModelElement.eClass)) a = A() assert a.eAnnotations == OrderedSet() annotation = EAnnotation(source='testAnnot') annotation.details['test'] = 'value' a.eAnnotations.append(annotation) assert len(a.eAnnotations) == 1 assert a.getEAnnotation('testAnnot') is annotation assert a.getEAnnotation('testAnnot').details['test'] == 'value' def test__EClass_extension(): SuperEClass = EClass('SuperEClass', superclass=(EClass.eClass,)) A = SuperEClass(name='A') assert isinstance(A, EClass) a = A() assert isinstance(a, EObject) assert a.eClass is A def test__EClass_modification(): EClass.new_feature = EAttribute('new_feature', EInt) A = EClass('A') assert A.new_feature == 0 A.new_feature = 5 assert A.new_feature == 5 with pytest.raises(BadValueError): A.new_feature = 'a'
Add tests for PyEcore extension (EClass/EModelElement tests)
Add tests for PyEcore extension (EClass/EModelElement tests)
Python
bsd-3-clause
aranega/pyecore,pyecore/pyecore
--- +++ @@ -0,0 +1,39 @@ +import pytest +from pyecore.ecore import * +import pyecore.ecore as ecore +from ordered_set import OrderedSet + + +def test__EModelElement_extension(): + A = EClass('A', superclass=(EModelElement.eClass)) + a = A() + assert a.eAnnotations == OrderedSet() + + annotation = EAnnotation(source='testAnnot') + annotation.details['test'] = 'value' + a.eAnnotations.append(annotation) + assert len(a.eAnnotations) == 1 + assert a.getEAnnotation('testAnnot') is annotation + assert a.getEAnnotation('testAnnot').details['test'] == 'value' + + +def test__EClass_extension(): + SuperEClass = EClass('SuperEClass', superclass=(EClass.eClass,)) + A = SuperEClass(name='A') + assert isinstance(A, EClass) + + a = A() + assert isinstance(a, EObject) + assert a.eClass is A + + +def test__EClass_modification(): + EClass.new_feature = EAttribute('new_feature', EInt) + A = EClass('A') + assert A.new_feature == 0 + + A.new_feature = 5 + assert A.new_feature == 5 + + with pytest.raises(BadValueError): + A.new_feature = 'a'
40431228c8535f325b005bb52485cae87a8be714
tests/unit/modules/test_napalm_acl.py
tests/unit/modules/test_napalm_acl.py
# -*- coding: utf-8 -*- ''' :codeauthor: :email:`Anthony Shaw <anthonyshaw@apache.org>` ''' # Import Python Libs from __future__ import absolute_import # Import Salt Testing Libs from tests.support.mixins import LoaderModuleMockMixin from tests.support.unit import TestCase, skipIf from tests.support.mock import ( MagicMock, NO_MOCK, NO_MOCK_REASON ) import tests.support.napalm as napalm_test_support import salt.modules.napalm_acl as napalm_acl # NOQA @skipIf(NO_MOCK, NO_MOCK_REASON) class NapalmAclModuleTestCase(TestCase, LoaderModuleMockMixin): def setup_loader_modules(self): module_globals = { '__salt__': { 'config.option': MagicMock(return_value={ 'test': { 'driver': 'test', 'key': '2orgk34kgk34g' } }), 'file.file_exists': napalm_test_support.true, 'file.join': napalm_test_support.join, 'file.get_managed': napalm_test_support.get_managed_file, 'random.hash': napalm_test_support.random_hash } } return {napalm_acl: module_globals} def test_load_term_config(self): ret = napalm_acl.load_term_config("test_filter", "test_term") assert ret['out'] is napalm_test_support.TEST_TERM_CONFIG
Add test module for napalm_acl
Add test module for napalm_acl
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
--- +++ @@ -0,0 +1,45 @@ +# -*- coding: utf-8 -*- +''' + :codeauthor: :email:`Anthony Shaw <anthonyshaw@apache.org>` +''' + +# Import Python Libs +from __future__ import absolute_import + +# Import Salt Testing Libs +from tests.support.mixins import LoaderModuleMockMixin +from tests.support.unit import TestCase, skipIf +from tests.support.mock import ( + MagicMock, + NO_MOCK, + NO_MOCK_REASON +) + +import tests.support.napalm as napalm_test_support +import salt.modules.napalm_acl as napalm_acl # NOQA + + +@skipIf(NO_MOCK, NO_MOCK_REASON) +class NapalmAclModuleTestCase(TestCase, LoaderModuleMockMixin): + + def setup_loader_modules(self): + module_globals = { + '__salt__': { + 'config.option': MagicMock(return_value={ + 'test': { + 'driver': 'test', + 'key': '2orgk34kgk34g' + } + }), + 'file.file_exists': napalm_test_support.true, + 'file.join': napalm_test_support.join, + 'file.get_managed': napalm_test_support.get_managed_file, + 'random.hash': napalm_test_support.random_hash + } + } + + return {napalm_acl: module_globals} + + def test_load_term_config(self): + ret = napalm_acl.load_term_config("test_filter", "test_term") + assert ret['out'] is napalm_test_support.TEST_TERM_CONFIG
841fb156fff3d257d39afdc9d3d4e587427fe2cf
Source/Scm/wb_scm_project_place_holder.py
Source/Scm/wb_scm_project_place_holder.py
''' ==================================================================== Copyright (c) 2016 Barry A Scott. All rights reserved. This software is licensed as described in the file LICENSE.txt, which you should have received as part of this distribution. ==================================================================== wb_scm_project_place_holder.py ''' import pathlib # # ScmProjectPlaceholder is used when the project cannot be loaded # class ScmProjectPlaceholder: def __init__( self, app, prefs_project ): self.app = app self.prefs_project = prefs_project self.tree = ScmProjectPlaceholderTreeNode( self, prefs_project.name, pathlib.Path( '.' ) ) def scmType( self ): return self.prefs_project.scm_type def isNotEqual( self, other ): return self.projectName() != other.projectName() def getBranchName( self ): return '' def projectName( self ): return self.prefs_project.name def projectPath( self ): return pathlib.Path( self.prefs_project.path ) def updateState( self ): pass class ScmProjectPlaceholderTreeNode: def __init__( self, project, name, path ): self.project = project self.name = name self.__path = path def __repr__( self ): return '<ScmProjectPlaceholderTreeNode: project %r, path %s>' % (self.project, self.__path) def isNotEqual( self, other ): return (self.relativePath() != other.relativePath() or self.project.isNotEqual( other.project )) def __lt__( self, other ): return self.name < other.name def relativePath( self ): return self.__path def absolutePath( self ): return self.project.projectPath() / self.__path def getAllFolderNodes( self ): return [] def getAllFolderNames( self ): return [] def getAllFileNames( self ): return [] def isByPath( self ): return False
Add new file missed in earlier commit place holder for projects that do not load for some reason
Add new file missed in earlier commit place holder for projects that do not load for some reason
Python
apache-2.0
barry-scott/git-workbench,barry-scott/scm-workbench,barry-scott/scm-workbench,barry-scott/git-workbench,barry-scott/scm-workbench
--- +++ @@ -0,0 +1,75 @@ +''' + ==================================================================== + Copyright (c) 2016 Barry A Scott. All rights reserved. + + This software is licensed as described in the file LICENSE.txt, + which you should have received as part of this distribution. + + ==================================================================== + + wb_scm_project_place_holder.py + +''' +import pathlib + +# +# ScmProjectPlaceholder is used when the project cannot be loaded +# +class ScmProjectPlaceholder: + def __init__( self, app, prefs_project ): + self.app = app + self.prefs_project = prefs_project + + self.tree = ScmProjectPlaceholderTreeNode( self, prefs_project.name, pathlib.Path( '.' ) ) + + def scmType( self ): + return self.prefs_project.scm_type + + def isNotEqual( self, other ): + return self.projectName() != other.projectName() + + def getBranchName( self ): + return '' + + def projectName( self ): + return self.prefs_project.name + + def projectPath( self ): + return pathlib.Path( self.prefs_project.path ) + + def updateState( self ): + pass + +class ScmProjectPlaceholderTreeNode: + def __init__( self, project, name, path ): + self.project = project + self.name = name + self.__path = path + + def __repr__( self ): + return '<ScmProjectPlaceholderTreeNode: project %r, path %s>' % (self.project, self.__path) + + def isNotEqual( self, other ): + return (self.relativePath() != other.relativePath() + or self.project.isNotEqual( other.project )) + + def __lt__( self, other ): + return self.name < other.name + + def relativePath( self ): + return self.__path + + def absolutePath( self ): + return self.project.projectPath() / self.__path + + def getAllFolderNodes( self ): + return [] + + def getAllFolderNames( self ): + return [] + + def getAllFileNames( self ): + return [] + + def isByPath( self ): + return False
5114f177741b105f33819b98415702e53b52eb01
corehq/apps/hqadmin/management/commands/update_site_setup.py
corehq/apps/hqadmin/management/commands/update_site_setup.py
from django.core.management.base import BaseCommand, CommandError from django.contrib.sites.models import Site from django.conf import settings class Command(BaseCommand): def add_arguments(self, parser): parser.add_argument( 'site_address', help="the new site address that should be used. This would get set in the site objects name " "and domain." ) parser.add_argument( '--skip-checks', action='store_true', default=False, help="If you are sure of what you are doing and want to skip checks to ensure safe update." ) def handle(self, site_address, *args, **options): if not options['skip_checks']: if settings.SITE_ID != 1: raise CommandError("SITE ID under settings expected to have value 1 since only one object is expected") sites_count = Site.objects.count() if sites_count != 1: raise CommandError("Expected to have only one object added by Site during setup but currently its %s "% Site.objects.count()) site_object = Site.objects.first() if site_object.name != "example.com" and site_object.domain != "example.com": raise CommandError( """ Expected the present site object to have dummy example values. They were probably modified and needs to be rechecked. Current Values, name -> {name}, domain -> {domain} """.format(name=site_object.name, domain=site_object.domain )) site_object = Site.objects.first() site_object.name = site_address site_object.domain = site_address site_object.save() Site.objects.clear_cache() site_object = Site.objects.first() print('Updated!') print('Site object now is name -> {name}, domain -> {domain}'.format( name=site_object.name, domain=site_object.domain ))
Add script to update site setup which is used at places like password reset email
Add script to update site setup which is used at places like password reset email [skip ci]
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
--- +++ @@ -0,0 +1,52 @@ +from django.core.management.base import BaseCommand, CommandError +from django.contrib.sites.models import Site +from django.conf import settings + + +class Command(BaseCommand): + def add_arguments(self, parser): + parser.add_argument( + 'site_address', + help="the new site address that should be used. This would get set in the site objects name " + "and domain." + ) + parser.add_argument( + '--skip-checks', + action='store_true', + default=False, + help="If you are sure of what you are doing and want to skip checks to ensure safe update." + ) + + def handle(self, site_address, *args, **options): + if not options['skip_checks']: + if settings.SITE_ID != 1: + raise CommandError("SITE ID under settings expected to have value 1 since only one object is expected") + sites_count = Site.objects.count() + if sites_count != 1: + raise CommandError("Expected to have only one object added by Site during setup but currently its %s "% + Site.objects.count()) + site_object = Site.objects.first() + if site_object.name != "example.com" and site_object.domain != "example.com": + raise CommandError( + """ + Expected the present site object to have dummy example values. + They were probably modified and needs to be rechecked. + Current Values, name -> {name}, domain -> {domain} + """.format(name=site_object.name, domain=site_object.domain + )) + + site_object = Site.objects.first() + site_object.name = site_address + site_object.domain = site_address + site_object.save() + + Site.objects.clear_cache() + + site_object = Site.objects.first() + print('Updated!') + print('Site object now is name -> {name}, domain -> {domain}'.format( + name=site_object.name, + domain=site_object.domain + )) + +
8b4c34e84d306b5f9021de47bc3ae9050e2fc2b3
compare_clouds.py
compare_clouds.py
#!/usr/bin/env python3 from pathlib import Path """Code for comparing point clouds""" cloud1Path = Path("./data/reconstructions/2016_10_24__17_43_17/reference.ply") cloud2Path = Path("./data/reconstructions/2016_10_24__17_43_17/high_quality.ply") from load_ply import load_ply cloud1PointData = load_ply(cloud1Path)[0][:,:3].copy() cloud2PointData = load_ply(cloud2Path)[0][:,:3].copy() #if __name__=='__main__': #pass
Fix loading of ply files exported by meshlab
Fix loading of ply files exported by meshlab
Python
mit
drewm1980/multi_view_stereo_benchmark,drewm1980/multi_view_stereo_benchmark,drewm1980/multi_view_stereo_benchmark
--- +++ @@ -0,0 +1,16 @@ +#!/usr/bin/env python3 + +from pathlib import Path + +"""Code for comparing point clouds""" + +cloud1Path = Path("./data/reconstructions/2016_10_24__17_43_17/reference.ply") +cloud2Path = Path("./data/reconstructions/2016_10_24__17_43_17/high_quality.ply") + +from load_ply import load_ply + +cloud1PointData = load_ply(cloud1Path)[0][:,:3].copy() +cloud2PointData = load_ply(cloud2Path)[0][:,:3].copy() + +#if __name__=='__main__': + #pass
0bf7d9fb20a3d2588ffc0e8341ec2af3df5fe300
depot/tests/test_depot_index.py
depot/tests/test_depot_index.py
from django.test import TestCase, Client from depot.models import Depot def create_depot(name, state): return Depot.objects.create(name=name, active=state) class DepotIndexTestCase(TestCase): def test_depot_index_template(self): response = self.client.get('/depots/') self.assertTemplateUsed( response, template_name='depot/index.html' ) def test_depot_index_with_no_depots(self): response = self.client.get('/depots/') self.assertEqual(response.status_code, 200) self.assertQuerysetEqual(response.context['depot_list'], []) self.assertContains(response, 'No depots available :(') def test_depot_index_with_active_depot(self): depot = create_depot('active depot', True) response = self.client.get('/depots/') self.assertEqual(response.status_code, 200) self.assertQuerysetEqual( response.context['depot_list'], ['<Depot: Depot active depot>'] ) self.assertContains(response, depot.name) def test_depot_index_with_archived_depot(self): depot = create_depot('archived depot', False) response = self.client.get('/depots/') self.assertEqual(response.status_code, 200) self.assertQuerysetEqual(response.context['depot_list'], []) self.assertContains(response, 'No depots available') self.assertNotContains(response, depot.name) def test_depot_index_with_active_and_archived_depot(self): active_depot = create_depot('active depot', True) archived_depot = create_depot('archived depot', False) response = self.client.get('/depots/') self.assertEqual(response.status_code, 200) self.assertQuerysetEqual( response.context['depot_list'], ['<Depot: Depot active depot>'] ) self.assertContains(response, active_depot.name) self.assertNotContains(response, archived_depot.name)
Add test for depot index page
Add test for depot index page
Python
agpl-3.0
verleihtool/verleihtool,verleihtool/verleihtool,verleihtool/verleihtool,verleihtool/verleihtool
--- +++ @@ -0,0 +1,50 @@ +from django.test import TestCase, Client +from depot.models import Depot + + +def create_depot(name, state): + return Depot.objects.create(name=name, active=state) + + +class DepotIndexTestCase(TestCase): + + def test_depot_index_template(self): + response = self.client.get('/depots/') + self.assertTemplateUsed( + response, + template_name='depot/index.html' + ) + + def test_depot_index_with_no_depots(self): + response = self.client.get('/depots/') + self.assertEqual(response.status_code, 200) + self.assertQuerysetEqual(response.context['depot_list'], []) + self.assertContains(response, 'No depots available :(') + + def test_depot_index_with_active_depot(self): + depot = create_depot('active depot', True) + response = self.client.get('/depots/') + self.assertEqual(response.status_code, 200) + self.assertQuerysetEqual( + response.context['depot_list'], ['<Depot: Depot active depot>'] + ) + self.assertContains(response, depot.name) + + def test_depot_index_with_archived_depot(self): + depot = create_depot('archived depot', False) + response = self.client.get('/depots/') + self.assertEqual(response.status_code, 200) + self.assertQuerysetEqual(response.context['depot_list'], []) + self.assertContains(response, 'No depots available') + self.assertNotContains(response, depot.name) + + def test_depot_index_with_active_and_archived_depot(self): + active_depot = create_depot('active depot', True) + archived_depot = create_depot('archived depot', False) + response = self.client.get('/depots/') + self.assertEqual(response.status_code, 200) + self.assertQuerysetEqual( + response.context['depot_list'], ['<Depot: Depot active depot>'] + ) + self.assertContains(response, active_depot.name) + self.assertNotContains(response, archived_depot.name)
cc79ee252e09ade17961d03265c61a87e270bd88
nototools/map_pua_emoji.py
nototools/map_pua_emoji.py
#!/usr/bin/python # # Copyright 2014 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Modify an emoji font to map legacy PUA characters to standard ligatures.""" __author__ = 'roozbeh@google.com (Roozbeh Pournader)' import sys from fontTools import ttLib from nototools import add_emoji_gsub from nototools import font_data def get_glyph_name_from_gsub(char_seq, font): """Find the glyph name for ligature of a given character sequence from GSUB. """ cmap = font_data.get_cmap(font) # FIXME: So many assumptions are made here. try: first_glyph = cmap[char_seq[0]] rest_of_glyphs = [cmap[ch] for ch in char_seq[1:]] except KeyError: return None for lookup in font['GSUB'].table.LookupList.Lookup: ligatures = lookup.SubTable[0].ligatures try: for ligature in ligatures[first_glyph]: if ligature.Component == rest_of_glyphs: return ligature.LigGlyph except KeyError: continue return None def add_pua_cmap(source_file, target_file): """Add PUA characters to the cmap of the first font and save as second.""" font = ttLib.TTFont(source_file) cmap = font_data.get_cmap(font) for pua, (ch1, ch2) in (add_emoji_gsub.EMOJI_KEYCAPS.items() + add_emoji_gsub.EMOJI_FLAGS.items()): if pua not in cmap: glyph_name = get_glyph_name_from_gsub([ch1, ch2], font) if glyph_name is not None: cmap[pua] = glyph_name font.save(target_file) def main(argv): """Save the first font given to the second font.""" add_pua_cmap(argv[1], argv[2]) if __name__ == '__main__': main(sys.argv)
Make color emoji use character sequences instead of PUA.
Make color emoji use character sequences instead of PUA. The bitmap emoji tools are extended to create GSUB rules for character sequences. The images are renamed to code the character sequence in their filenames. New images are created for ASCII digits and number sign. A new script is added to add cmaps to the files for backward compatibility.
Python
apache-2.0
davelab6/nototools,anthrotype/nototools,pathumego/nototools,googlei18n/nototools,davelab6/nototools,pahans/nototools,moyogo/nototools,googlefonts/nototools,namemealrady/nototools,googlei18n/nototools,moyogo/nototools,dougfelt/nototools,googlefonts/nototools,pahans/nototools,googlei18n/nototools,googlefonts/nototools,davelab6/nototools,namemealrady/nototools,dougfelt/nototools,pathumego/nototools,anthrotype/nototools,dougfelt/nototools,googlefonts/nototools,anthrotype/nototools,pahans/nototools,namemealrady/nototools,pathumego/nototools,googlefonts/nototools,moyogo/nototools
--- +++ @@ -0,0 +1,71 @@ +#!/usr/bin/python +# +# Copyright 2014 Google Inc. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""Modify an emoji font to map legacy PUA characters to standard ligatures.""" + +__author__ = 'roozbeh@google.com (Roozbeh Pournader)' + +import sys + +from fontTools import ttLib + +from nototools import add_emoji_gsub +from nototools import font_data + + +def get_glyph_name_from_gsub(char_seq, font): + """Find the glyph name for ligature of a given character sequence from GSUB. + """ + cmap = font_data.get_cmap(font) + # FIXME: So many assumptions are made here. + try: + first_glyph = cmap[char_seq[0]] + rest_of_glyphs = [cmap[ch] for ch in char_seq[1:]] + except KeyError: + return None + + for lookup in font['GSUB'].table.LookupList.Lookup: + ligatures = lookup.SubTable[0].ligatures + try: + for ligature in ligatures[first_glyph]: + if ligature.Component == rest_of_glyphs: + return ligature.LigGlyph + except KeyError: + continue + return None + + +def add_pua_cmap(source_file, target_file): + """Add PUA characters to the cmap of the first font and save as second.""" + font = ttLib.TTFont(source_file) + cmap = font_data.get_cmap(font) + for pua, (ch1, ch2) in (add_emoji_gsub.EMOJI_KEYCAPS.items() + + add_emoji_gsub.EMOJI_FLAGS.items()): + if pua not in cmap: + glyph_name = get_glyph_name_from_gsub([ch1, ch2], font) + if glyph_name is not None: + cmap[pua] = glyph_name + font.save(target_file) + + +def main(argv): + """Save the first font given to the second font.""" + add_pua_cmap(argv[1], argv[2]) + + +if __name__ == '__main__': + main(sys.argv) +
3adcefcad4fc3ecb85aa4a22e8b3c4bf5ca4e6f5
test/integration/ggrc/converters/test_import_update.py
test/integration/ggrc/converters/test_import_update.py
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from integration.ggrc.converters import TestCase from ggrc import models class TestImportUpdates(TestCase): """ Test importing of already existing objects """ def setUp(self): TestCase.setUp(self) self.client.get("/login") def test_policy_basic_update(self): """ Test simple policy title update """ filename = "policy_basic_import.csv" response = self.import_file(filename) self._check_response(response, {}) policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "some weird policy") filename = "policy_basic_import_update.csv" response = self.import_file(filename) self._check_response(response, {}) policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "Edited policy")
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> """Tests for bulk updates with CSV import.""" from integration.ggrc.converters import TestCase from ggrc import models class TestImportUpdates(TestCase): """ Test importing of already existing objects """ def setUp(self): TestCase.setUp(self) self.client.get("/login") def test_policy_basic_update(self): """ Test simple policy title update """ filename = "policy_basic_import.csv" response = self.import_file(filename) self._check_response(response, {}) policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "some weird policy") revision_count = models.Revision.query.filter( models.Revision.resource_type == "Policy", models.Revision.resource_id == policy.id ).count() self.assertEqual(revision_count, 1) filename = "policy_basic_import_update.csv" response = self.import_file(filename) self._check_response(response, {}) policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "Edited policy") revision_count = models.Revision.query.filter( models.Revision.resource_type == "Policy", models.Revision.resource_id == policy.id ).count() self.assertEqual(revision_count, 2)
Add tests for revision updates via import
Add tests for revision updates via import This tests checks if new revisions were added for object updated via CSV import.
Python
apache-2.0
edofic/ggrc-core,selahssea/ggrc-core,edofic/ggrc-core,plamut/ggrc-core,josthkko/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,andrei-karalionak/ggrc-core,AleksNeStu/ggrc-core,selahssea/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,andrei-karalionak/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,edofic/ggrc-core,josthkko/ggrc-core,andrei-karalionak/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,selahssea/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,AleksNeStu/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,plamut/ggrc-core,josthkko/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core,kr41/ggrc-core,edofic/ggrc-core,VinnieJohns/ggrc-core
--- +++ @@ -1,5 +1,7 @@ # Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> + +"""Tests for bulk updates with CSV import.""" from integration.ggrc.converters import TestCase @@ -24,6 +26,11 @@ policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "some weird policy") + revision_count = models.Revision.query.filter( + models.Revision.resource_type == "Policy", + models.Revision.resource_id == policy.id + ).count() + self.assertEqual(revision_count, 1) filename = "policy_basic_import_update.csv" response = self.import_file(filename) @@ -31,3 +38,8 @@ policy = models.Policy.query.filter_by(slug="p1").first() self.assertEqual(policy.title, "Edited policy") + revision_count = models.Revision.query.filter( + models.Revision.resource_type == "Policy", + models.Revision.resource_id == policy.id + ).count() + self.assertEqual(revision_count, 2)
e48caa4bb61cce466ad5eb9bffbfba8e33312474
python/example_code/ec2/terminate_instances.py
python/example_code/ec2/terminate_instances.py
# snippet-comment:[These are tags for the AWS doc team's sample catalog. Do not remove.] # snippet-sourcedescription:[terminate_instances.py demonstrates how to terminate an Amazon EC2 instance.] # snippet-service:[ec2] # snippet-keyword:[Amazon EC2] # snippet-keyword:[Python] # snippet-keyword:[AWS SDK for Python (Boto3)] # snippet-keyword:[Code Sample] # snippet-sourcetype:[full-example] # snippet-sourcedate:[2019-2-11] # snippet-sourceauthor:[AWS] # Copyright 2010-2019 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # This file is licensed under the Apache License, Version 2.0 (the "License"). # You may not use this file except in compliance with the License. A copy of the # License is located at # # http://aws.amazon.com/apache2.0/ # # This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS # OF ANY KIND, either express or implied. See the License for the specific # language governing permissions and limitations under the License. import logging import boto3 from botocore.exceptions import ClientError def terminate_instances(instance_ids): """Terminate one or more Amazon EC2 instances :param instance_ids: List of string IDs of EC2 instances to terminate :return: List of state information for each instance specified in instance_ids. If error, return None. """ # Terminate each instance in the argument list ec2 = boto3.client('ec2') try: states = ec2.terminate_instances(InstanceIds=instance_ids) except ClientError as e: logging.error(e) return None return states['TerminatingInstances'] def main(): """Exercise terminate_instances()""" # Assign these values before running the program ec2_instance_ids = ['EC2_INSTANCE_ID'] # Set up logging logging.basicConfig(level=logging.DEBUG, format='%(levelname)s: %(asctime)s: %(message)s') # Terminate the EC2 instance(s) states = terminate_instances(ec2_instance_ids) if states is not None: logging.debug('Terminating the following EC2 instances') for state in states: logging.debug(f'ID: {state["InstanceId"]}') logging.debug(f' Current state: Code {state["CurrentState"]["Code"]}, ' f'{state["CurrentState"]["Name"]}') logging.debug(f' Previous state: Code {state["PreviousState"]["Code"]}, ' f'{state["PreviousState"]["Name"]}') if __name__ == '__main__': main()
Add Python EC2 TerminateInstances example
Add Python EC2 TerminateInstances example Merge pull request #504 from awsdocs/scalwas_ec2
Python
apache-2.0
awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples
--- +++ @@ -0,0 +1,71 @@ +# snippet-comment:[These are tags for the AWS doc team's sample catalog. Do not remove.] +# snippet-sourcedescription:[terminate_instances.py demonstrates how to terminate an Amazon EC2 instance.] +# snippet-service:[ec2] +# snippet-keyword:[Amazon EC2] +# snippet-keyword:[Python] +# snippet-keyword:[AWS SDK for Python (Boto3)] +# snippet-keyword:[Code Sample] +# snippet-sourcetype:[full-example] +# snippet-sourcedate:[2019-2-11] +# snippet-sourceauthor:[AWS] + +# Copyright 2010-2019 Amazon.com, Inc. or its affiliates. All Rights Reserved. +# +# This file is licensed under the Apache License, Version 2.0 (the "License"). +# You may not use this file except in compliance with the License. A copy of the +# License is located at +# +# http://aws.amazon.com/apache2.0/ +# +# This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS +# OF ANY KIND, either express or implied. See the License for the specific +# language governing permissions and limitations under the License. + +import logging +import boto3 +from botocore.exceptions import ClientError + + +def terminate_instances(instance_ids): + """Terminate one or more Amazon EC2 instances + + :param instance_ids: List of string IDs of EC2 instances to terminate + :return: List of state information for each instance specified in instance_ids. + If error, return None. + """ + + + # Terminate each instance in the argument list + ec2 = boto3.client('ec2') + try: + states = ec2.terminate_instances(InstanceIds=instance_ids) + except ClientError as e: + logging.error(e) + return None + return states['TerminatingInstances'] + + +def main(): + """Exercise terminate_instances()""" + + # Assign these values before running the program + ec2_instance_ids = ['EC2_INSTANCE_ID'] + + # Set up logging + logging.basicConfig(level=logging.DEBUG, + format='%(levelname)s: %(asctime)s: %(message)s') + + # Terminate the EC2 instance(s) + states = terminate_instances(ec2_instance_ids) + if states is not None: + logging.debug('Terminating the following EC2 instances') + for state in states: + logging.debug(f'ID: {state["InstanceId"]}') + logging.debug(f' Current state: Code {state["CurrentState"]["Code"]}, ' + f'{state["CurrentState"]["Name"]}') + logging.debug(f' Previous state: Code {state["PreviousState"]["Code"]}, ' + f'{state["PreviousState"]["Name"]}') + + +if __name__ == '__main__': + main()
58fee826ab5298f7de036bf320bbc109b853eec8
tendrl/commons/manager/__init__.py
tendrl/commons/manager/__init__.py
import abc import logging import six from tendrl.commons import jobs LOG = logging.getLogger(__name__) @six.add_metaclass(abc.ABCMeta) class Manager(object): def __init__( self, sds_sync_thread, central_store_thread, ): self._central_store_thread = central_store_thread self._sds_sync_thread = sds_sync_thread self._job_consumer_thread = jobs.JobConsumerThread() def stop(self): LOG.info("%s stopping" % self.__class__.__name__) self._job_consumer_thread.stop() self._sds_sync_thread.stop() self._central_store_thread.stop() def start(self): LOG.info("%s starting" % self.__class__.__name__) self._central_store_thread.start() self._sds_sync_thread.start() self._job_consumer_thread.start() def join(self): LOG.info("%s joining" % self.__class__.__name__) self._job_consumer_thread.join() self._sds_sync_thread.join() self._central_store_thread.join()
import abc import logging import six from tendrl.commons import jobs LOG = logging.getLogger(__name__) @six.add_metaclass(abc.ABCMeta) class Manager(object): def __init__( self, sds_sync_thread, central_store_thread, ): self._central_store_thread = central_store_thread self._sds_sync_thread = sds_sync_thread self._job_consumer_thread = jobs.JobConsumerThread() def stop(self): LOG.info("%s stopping" % self.__class__.__name__) self._job_consumer_thread.stop() if self._sds_sync_thread: self._sds_sync_thread.stop() self._central_store_thread.stop() def start(self): LOG.info("%s starting" % self.__class__.__name__) self._central_store_thread.start() if self._sds_sync_thread: self._sds_sync_thread.start() self._job_consumer_thread.start() def join(self): LOG.info("%s joining" % self.__class__.__name__) self._job_consumer_thread.join() if self._sds_sync_thread: self._sds_sync_thread.join() self._central_store_thread.join()
Add null check for sds sync thread which can be optional
Add null check for sds sync thread which can be optional Signed-off-by: anmolbabu <3d38fb1e9c5ff2662fc415254efcdfedb95b84d5@gmail.com>
Python
lgpl-2.1
Tendrl/commons,rishubhjain/commons,r0h4n/commons
--- +++ @@ -23,19 +23,21 @@ def stop(self): LOG.info("%s stopping" % self.__class__.__name__) self._job_consumer_thread.stop() - self._sds_sync_thread.stop() + if self._sds_sync_thread: + self._sds_sync_thread.stop() self._central_store_thread.stop() def start(self): LOG.info("%s starting" % self.__class__.__name__) self._central_store_thread.start() - self._sds_sync_thread.start() + if self._sds_sync_thread: + self._sds_sync_thread.start() self._job_consumer_thread.start() def join(self): LOG.info("%s joining" % self.__class__.__name__) self._job_consumer_thread.join() - self._sds_sync_thread.join() + if self._sds_sync_thread: + self._sds_sync_thread.join() self._central_store_thread.join() -
e5a39d4e17a0555cb242731b34f0ee480367b4fe
foireminder/foireminder/reminders/tasks.py
foireminder/foireminder/reminders/tasks.py
from django.utils import timezone from .models import ReminderRequest, EmailReminder def send_todays_notifications(self): today = timezone.now() reminders = ReminderRequest.objects.filter( start__year=today.year, start__month=today.month, start__day=today.da ) for reminder in reminders: for subscriber in EmailReminder.objects.filter(rule=reminder.rule): subscriber.send_notification()
Add task that sends out notifications
Add task that sends out notifications
Python
mit
stefanw/foireminder,stefanw/foireminder
--- +++ @@ -0,0 +1,15 @@ +from django.utils import timezone + +from .models import ReminderRequest, EmailReminder + + +def send_todays_notifications(self): + today = timezone.now() + reminders = ReminderRequest.objects.filter( + start__year=today.year, + start__month=today.month, + start__day=today.da + ) + for reminder in reminders: + for subscriber in EmailReminder.objects.filter(rule=reminder.rule): + subscriber.send_notification()
c58c58d5bf1394e04e30f5eeb298818558be027f
tests/rules_tests/clearAfterNonTermRemove/__init__.py
tests/rules_tests/clearAfterNonTermRemove/__init__.py
#!/usr/bin/env python """ :Author Patrik Valkovic :Created 17.08.2017 22:06 :Licence GNUv3 Part of grammpy """
Add directory for tests of rules removin
Add directory for tests of rules removin
Python
mit
PatrikValkovic/grammpy
--- +++ @@ -0,0 +1,8 @@ +#!/usr/bin/env python +""" +:Author Patrik Valkovic +:Created 17.08.2017 22:06 +:Licence GNUv3 +Part of grammpy + +"""
d40fb122d7083b9735728df15120ed682431be79
scripts/make_fhes_seeds.py
scripts/make_fhes_seeds.py
import yaml import sys import numpy as np from astropy.table import Table from astropy.coordinates import SkyCoord from fermipy.catalog import * from fermipy.utils import * def get_coord(name,tab): row = tab[tab['Source_Name'] == name] return SkyCoord(float(row['RAJ2000']), float(row['DEJ2000']),unit='deg') def avg_coords(coords): xyz = np.zeros(3) for t in coords: xyz += t.cartesian.xyz xyz /= np.sum(xyz**2)**0.5 c = SkyCoord(xyz[0], xyz[1], xyz[2],representation='cartesian') c.representation='spherical' return c tab = Table.read(sys.argv[1]) src_names = [] m = np.abs(tab['glat']) < 0. #m |= (tab['fit_ext_gauss_ts_ext'] > 9.0) #m |= (tab['fit_ext_disk_ts_ext'] > 9.0) m |= (tab['fit_halo_ts'] > 16.0) #m |= (tab['ts'] > 20000.0) for row in tab[m]: src_names += [row['codename']] src_names = sorted(list(set(src_names))) o = {} for name in src_names: #coords = [get_coord(t,cat.table) for t in names] #c0 = avg_coords(coords) print(name) #print(create_source_name(c0)) names = [name] row = tab[tab['codename'] == names[0].lower().replace(' ','_')] c0 = SkyCoord(row['ra'],row['dec'],unit='deg') name = create_source_name(c0).replace('PS','FHES') + 'e' #print(c0.ra.deg,c0.dec.deg) #print(names[0]) #print(row['codename']) src = {'name' : name, 'ra' : float(c0.ra.deg), 'dec' : float(c0.dec.deg), 'SpectrumType' : 'PowerLaw', 'SpatialModel' : 'RadialGaussian', 'SpatialWidth' : float(row['fit_halo_r68']), 'Index' : float(row['fit_halo_index'])} o[name.lower().replace(' ','_')] = {'selection' : {'target' : name}, 'model' : {'sources' : [src]} } yaml.dump(o,open('out.yaml','w'))
Create script for generating analysis seeds.
Create script for generating analysis seeds.
Python
bsd-3-clause
woodmd/haloanalysis,woodmd/haloanalysis
--- +++ @@ -0,0 +1,72 @@ +import yaml +import sys +import numpy as np +from astropy.table import Table +from astropy.coordinates import SkyCoord +from fermipy.catalog import * +from fermipy.utils import * + + +def get_coord(name,tab): + + row = tab[tab['Source_Name'] == name] + return SkyCoord(float(row['RAJ2000']), float(row['DEJ2000']),unit='deg') + +def avg_coords(coords): + xyz = np.zeros(3) + for t in coords: + xyz += t.cartesian.xyz + + xyz /= np.sum(xyz**2)**0.5 + + c = SkyCoord(xyz[0], xyz[1], xyz[2],representation='cartesian') + c.representation='spherical' + return c + + + +tab = Table.read(sys.argv[1]) +src_names = [] + +m = np.abs(tab['glat']) < 0. +#m |= (tab['fit_ext_gauss_ts_ext'] > 9.0) +#m |= (tab['fit_ext_disk_ts_ext'] > 9.0) +m |= (tab['fit_halo_ts'] > 16.0) +#m |= (tab['ts'] > 20000.0) + + +for row in tab[m]: + src_names += [row['codename']] + +src_names = sorted(list(set(src_names))) + +o = {} + +for name in src_names: + #coords = [get_coord(t,cat.table) for t in names] + #c0 = avg_coords(coords) + + print(name) + #print(create_source_name(c0)) + + names = [name] + + row = tab[tab['codename'] == names[0].lower().replace(' ','_')] + c0 = SkyCoord(row['ra'],row['dec'],unit='deg') + name = create_source_name(c0).replace('PS','FHES') + 'e' + #print(c0.ra.deg,c0.dec.deg) + + #print(names[0]) + #print(row['codename']) + + src = {'name' : name, + 'ra' : float(c0.ra.deg), 'dec' : float(c0.dec.deg), + 'SpectrumType' : 'PowerLaw', 'SpatialModel' : 'RadialGaussian', + 'SpatialWidth' : float(row['fit_halo_r68']), + 'Index' : float(row['fit_halo_index'])} + + o[name.lower().replace(' ','_')] = {'selection' : {'target' : name}, + 'model' : {'sources' : [src]} } + + +yaml.dump(o,open('out.yaml','w'))
d1ee86414d45c571571d75434b8c2256b0120732
py/binary-tree-tilt.py
py/binary-tree-tilt.py
# Definition for a binary tree node. # class TreeNode(object): # def __init__(self, x): # self.val = x # self.left = None # self.right = None class Solution(object): def findTilt(self, root): """ :type root: TreeNode :rtype: int """ return self.do_findTilt(root)[1] def do_findTilt(self, cur): if cur is None: return (0, 0) lsum, ltiltsum = self.do_findTilt(cur.left) rsum, rtiltsum = self.do_findTilt(cur.right) tilt = abs(lsum - rsum) return lsum + rsum + cur.val, ltiltsum + rtiltsum + tilt
Add py solution for 563. Binary Tree Tilt
Add py solution for 563. Binary Tree Tilt 563. Binary Tree Tilt: https://leetcode.com/problems/binary-tree-tilt/
Python
apache-2.0
ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode
--- +++ @@ -0,0 +1,22 @@ +# Definition for a binary tree node. +# class TreeNode(object): +# def __init__(self, x): +# self.val = x +# self.left = None +# self.right = None + +class Solution(object): + def findTilt(self, root): + """ + :type root: TreeNode + :rtype: int + """ + return self.do_findTilt(root)[1] + + def do_findTilt(self, cur): + if cur is None: + return (0, 0) + lsum, ltiltsum = self.do_findTilt(cur.left) + rsum, rtiltsum = self.do_findTilt(cur.right) + tilt = abs(lsum - rsum) + return lsum + rsum + cur.val, ltiltsum + rtiltsum + tilt
925ff38344b5058ce196877e1fdcf79a1d1f6719
ue4/tests/test_messaging.py
ue4/tests/test_messaging.py
import pytest from m2u.ue4 import connection def test_send_message_size(): """Send a big message, larger than buffer size, so the server has to read multiple chunks. """ message = "TestMessageSize " + ("abcdefg" * 5000) connection.connect() result = connection.send_message(message) assert result == str(len(message)) connection.disconnect()
Add basic test for checking messages are received correctly
Add basic test for checking messages are received correctly
Python
mit
m2u/m2u
--- +++ @@ -0,0 +1,15 @@ +import pytest + +from m2u.ue4 import connection + + +def test_send_message_size(): + """Send a big message, larger than buffer size, so the server has to + read multiple chunks. + + """ + message = "TestMessageSize " + ("abcdefg" * 5000) + connection.connect() + result = connection.send_message(message) + assert result == str(len(message)) + connection.disconnect()
ff700e5d6fc5e0c5062f687110563d7f0312a3f0
server/tests/test_admin.py
server/tests/test_admin.py
"""General functional tests for the API endpoints.""" from django.test import TestCase, Client # from django.urls import reverse from rest_framework import status from server.models import ApiKey, User # from api.v2.tests.tools import SalAPITestCase class AdminTest(TestCase): """Test the admin site is configured to have all expected views.""" admin_endpoints = { 'apikey', 'businessunit', 'condition', 'fact', 'historicalfact', 'installedupdate', 'machinedetailplugin', 'machinegroup', 'machine', 'pendingappleupdate', 'pendingupdate', 'pluginscriptrow', 'pluginscriptsubmission', 'plugin', 'report', 'salsetting', 'updatehistoryitem', 'updatehistory', 'userprofile'} def setUp(self): self.client = Client() self.user = User.objects.create(username='test') def test_no_access(self): """Test that unauthenticated requests redirected to login.""" for path in self.admin_endpoints: response = self.client.get('/admin/server/{}'.format(path)) # Redirect to login page. self.assertEqual(response.status_code, status.HTTP_301_MOVED_PERMANENTLY) def test_ro_access(self): """Test that ro requests are rejected. RO users should not have access to the admin site (unless they have `is_staff = True`. """ self.user.user_profile = 'RO' self.user.save() self.client.force_login(self.user) for path in self.admin_endpoints: url = '/admin/server/{}/'.format(path) response = self.client.get(url) msg = 'Failed for path: "{}"'.format(path) self.assertEqual(response.status_code, status.HTTP_302_FOUND, msg=msg) self.assertEqual(response.url, '/admin/login/?next=/admin/server/{}/'.format(path), msg=msg) def test_ga_access(self): """Ensure GA userprofile grants admin page access.""" self.user.user_profile = 'GA' self.user.save() self.client.force_login(self.user) for path in self.admin_endpoints: url = '/admin/server/{}/'.format(path) response = self.client.get(url, follow=True) msg = 'Failed for path: "{}"'.format(path) self.assertEqual(response.status_code, status.HTTP_200_OK, msg=msg)
Set up test suite to ensure server admin routes are added.
Set up test suite to ensure server admin routes are added.
Python
apache-2.0
sheagcraig/sal,salopensource/sal,salopensource/sal,sheagcraig/sal,salopensource/sal,sheagcraig/sal,sheagcraig/sal,salopensource/sal
--- +++ @@ -0,0 +1,61 @@ +"""General functional tests for the API endpoints.""" + + +from django.test import TestCase, Client +# from django.urls import reverse + +from rest_framework import status + +from server.models import ApiKey, User +# from api.v2.tests.tools import SalAPITestCase + + +class AdminTest(TestCase): + """Test the admin site is configured to have all expected views.""" + admin_endpoints = { + 'apikey', 'businessunit', 'condition', 'fact', 'historicalfact', + 'installedupdate', 'machinedetailplugin', 'machinegroup', 'machine', + 'pendingappleupdate', 'pendingupdate', 'pluginscriptrow', + 'pluginscriptsubmission', 'plugin', 'report', 'salsetting', 'updatehistoryitem', + 'updatehistory', 'userprofile'} + + def setUp(self): + self.client = Client() + self.user = User.objects.create(username='test') + + def test_no_access(self): + """Test that unauthenticated requests redirected to login.""" + for path in self.admin_endpoints: + response = self.client.get('/admin/server/{}'.format(path)) + # Redirect to login page. + self.assertEqual(response.status_code, status.HTTP_301_MOVED_PERMANENTLY) + + def test_ro_access(self): + """Test that ro requests are rejected. + + RO users should not have access to the admin site (unless they have + `is_staff = True`. + """ + self.user.user_profile = 'RO' + self.user.save() + self.client.force_login(self.user) + + for path in self.admin_endpoints: + url = '/admin/server/{}/'.format(path) + response = self.client.get(url) + msg = 'Failed for path: "{}"'.format(path) + self.assertEqual(response.status_code, status.HTTP_302_FOUND, msg=msg) + self.assertEqual(response.url, '/admin/login/?next=/admin/server/{}/'.format(path), + msg=msg) + + def test_ga_access(self): + """Ensure GA userprofile grants admin page access.""" + self.user.user_profile = 'GA' + self.user.save() + self.client.force_login(self.user) + + for path in self.admin_endpoints: + url = '/admin/server/{}/'.format(path) + response = self.client.get(url, follow=True) + msg = 'Failed for path: "{}"'.format(path) + self.assertEqual(response.status_code, status.HTTP_200_OK, msg=msg)
0c29b431a0f5ce9115d7acdcaaabbd27546949c6
chmvh_website/contact/tests/views/test_success_view.py
chmvh_website/contact/tests/views/test_success_view.py
from django.test import RequestFactory from django.urls import reverse from contact.views import SuccessView class TestSuccessView(object): """Test cases for the success view""" url = reverse('contact:success') def test_get(self, rf: RequestFactory): """Test sending a GET request to the view. Sending a GET request to the view should render the success page. """ request = rf.get(self.url) response = SuccessView.as_view()(request) assert response.status_code == 200 assert 'contact/success.html' in response.template_name
Add test for contact success view.
Add test for contact success view.
Python
mit
cdriehuys/chmvh-website,cdriehuys/chmvh-website,cdriehuys/chmvh-website
--- +++ @@ -0,0 +1,21 @@ +from django.test import RequestFactory +from django.urls import reverse + +from contact.views import SuccessView + + +class TestSuccessView(object): + """Test cases for the success view""" + url = reverse('contact:success') + + def test_get(self, rf: RequestFactory): + """Test sending a GET request to the view. + + Sending a GET request to the view should render the success + page. + """ + request = rf.get(self.url) + response = SuccessView.as_view()(request) + + assert response.status_code == 200 + assert 'contact/success.html' in response.template_name
c650d64247d63d2af7a8168795e7edae5c9ef6ef
realtime-plot.py
realtime-plot.py
import time, random import math from collections import deque start = time.time() class RealtimePlot: def __init__(self, axes, max_entries = 100): self.axis_x = deque(maxlen=max_entries) self.axis_y = deque(maxlen=max_entries) self.axes = axes self.max_entries = max_entries self.lineplot, = axes.plot([], [], "ro-") self.axes.set_autoscaley_on(True) def add(self, x, y): self.axis_x.append(x) self.axis_y.append(y) self.lineplot.set_data(self.axis_x, self.axis_y) self.axes.set_xlim(self.axis_x[0], self.axis_x[-1] + 1e-15) self.axes.relim(); self.axes.autoscale_view() # rescale the y-axis def animate(self, figure, callback, interval = 50): import matplotlib.animation as animation def wrapper(frame_index): self.add(*callback(frame_index)) self.axes.relim(); self.axes.autoscale_view() # rescale the y-axis return self.lineplot animation.FuncAnimation(figure, wrapper, interval=interval) def main(): from matplotlib import pyplot as plt fig, axes = plt.subplots() display = RealtimePlot(axes) display.animate(fig, lambda frame_index: (time.time() - start, random.random() * 100)) plt.show() fig, axes = plt.subplots() display = RealtimePlot(axes) while True: display.add(time.time() - start, random.random() * 100) plt.pause(0.001) if __name__ == "__main__": main()
Add realtime chart plotting example
Add realtime chart plotting example
Python
mit
voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts
--- +++ @@ -0,0 +1,46 @@ +import time, random +import math +from collections import deque + +start = time.time() + +class RealtimePlot: + def __init__(self, axes, max_entries = 100): + self.axis_x = deque(maxlen=max_entries) + self.axis_y = deque(maxlen=max_entries) + self.axes = axes + self.max_entries = max_entries + + self.lineplot, = axes.plot([], [], "ro-") + self.axes.set_autoscaley_on(True) + + def add(self, x, y): + self.axis_x.append(x) + self.axis_y.append(y) + self.lineplot.set_data(self.axis_x, self.axis_y) + self.axes.set_xlim(self.axis_x[0], self.axis_x[-1] + 1e-15) + self.axes.relim(); self.axes.autoscale_view() # rescale the y-axis + + def animate(self, figure, callback, interval = 50): + import matplotlib.animation as animation + def wrapper(frame_index): + self.add(*callback(frame_index)) + self.axes.relim(); self.axes.autoscale_view() # rescale the y-axis + return self.lineplot + animation.FuncAnimation(figure, wrapper, interval=interval) + +def main(): + from matplotlib import pyplot as plt + + fig, axes = plt.subplots() + display = RealtimePlot(axes) + display.animate(fig, lambda frame_index: (time.time() - start, random.random() * 100)) + plt.show() + + fig, axes = plt.subplots() + display = RealtimePlot(axes) + while True: + display.add(time.time() - start, random.random() * 100) + plt.pause(0.001) + +if __name__ == "__main__": main()
f7e4ca11c7bfc35bf0fd6becd2a5d5fdd2ca5ed5
src/main/python/partition_data.py
src/main/python/partition_data.py
import csv; import random; import sys; in_file = str(sys.argv[1]) out_file = str(sys.argv[2]) num_partitions = int(sys.argv[3]) header = []; partitions = []; for i in range(num_partitions): partitions.append([]) # Load all the training rows row_num = 0; with open(in_file) as file: reader = csv.reader(file); header = reader.next(); for row in reader: partitions[row_num % num_partitions].append(row); row_num += 1; # Write test and train files for k partitions for i in range(num_partitions): train_rows = [] test_rows = partitions[i]; for j in range(num_partitions): if i != j: for row in partitions[j]: train_rows.append(row); with open(out_file+'_k'+str(i+1)+'_train.csv', 'wb') as ofile: writer = csv.writer(ofile) writer.writerow(header) for row in train_rows: writer.writerow(row) with open(out_file+'_k'+str(i+1)+'_test.csv', 'wb') as ofile: writer = csv.writer(ofile) writer.writerow(header) for row in test_rows: writer.writerow(row)
Add a script to split data with partitions.
Add a script to split data with partitions.
Python
mit
juckele/ddr-grader,juckele/ddr-grader
--- +++ @@ -0,0 +1,42 @@ +import csv; +import random; +import sys; + +in_file = str(sys.argv[1]) +out_file = str(sys.argv[2]) +num_partitions = int(sys.argv[3]) + +header = []; +partitions = []; +for i in range(num_partitions): + partitions.append([]) + +# Load all the training rows +row_num = 0; +with open(in_file) as file: + reader = csv.reader(file); + header = reader.next(); + for row in reader: + partitions[row_num % num_partitions].append(row); + row_num += 1; + +# Write test and train files for k partitions +for i in range(num_partitions): + train_rows = [] + test_rows = partitions[i]; + for j in range(num_partitions): + if i != j: + for row in partitions[j]: + train_rows.append(row); + + with open(out_file+'_k'+str(i+1)+'_train.csv', 'wb') as ofile: + writer = csv.writer(ofile) + writer.writerow(header) + for row in train_rows: + writer.writerow(row) + + with open(out_file+'_k'+str(i+1)+'_test.csv', 'wb') as ofile: + writer = csv.writer(ofile) + writer.writerow(header) + for row in test_rows: + writer.writerow(row)
645507ed9ec43b354880673fbc75afe169ef6697
tests/unit/test_handlers.py
tests/unit/test_handlers.py
from pmxbot import core def test_contains_always_match(): """ Contains handler should always match if no rate is specified. """ handler = core.ContainsHandler(name='#', func=None) assert handler.match('Tell me about #foo', channel='bar')
Add test capturing bad implementation of contains handler.
Add test capturing bad implementation of contains handler.
Python
bsd-3-clause
jawilson/pmxbot,jawilson/pmxbot
--- +++ @@ -0,0 +1,8 @@ +from pmxbot import core + +def test_contains_always_match(): + """ + Contains handler should always match if no rate is specified. + """ + handler = core.ContainsHandler(name='#', func=None) + assert handler.match('Tell me about #foo', channel='bar')
b3e9075e819402f93f7dc2e29b61e3e621ab7355
impy/imputations/tests/test_averaging_imputations.py
impy/imputations/tests/test_averaging_imputations.py
"""test_averaging_imputations.py""" import unittest import numpy as np from impy.imputations import mean_imputation from impy.imputations import mode_imputation from impy.imputations import median_imputation from impy.datasets import random_int class TestAveraging(unittest.TestCase): """ Tests for Averaging """ def setUp(self): self.data = random_int(missingness="complete") def test_mean_return_type(self): """Mean Imputation Return Type""" self.assertEqual(str(type(mean_imputation(self.data))), "<class 'numpy.ndarray'>") def test_mode_return_type(self): """Mode Imputation Return Type""" self.assertEqual(str(type(mode_imputation(self.data))), "<class 'numpy.ndarray'>") def test_median_return_type(self): """Median Imputation Return Type""" self.assertEqual(str(type(median_imputation(self.data))), "<class 'numpy.ndarray'>") def test_mean_fill(self): """ Mean Imputation Fill Complete Data(nothing should happen)""" actual = mean_imputation(self.data) self.assertTrue(np.array_equal(actual, self.data)) def test_mode_fill(self): """ Mode Imputation Fill Complete Data(nothing should happen)""" actual = mode_imputation(self.data) self.assertTrue(np.array_equal(actual, self.data)) def test_median_fill(self): """ Median Imputation Fill Complete Data(nothing should happen)""" actual = median_imputation(self.data) self.assertTrue(np.array_equal(actual, self.data)) if __name__ == "__main__": unittest.main()
Add unit tests for avging imputations
Add unit tests for avging imputations
Python
mit
eltonlaw/impyute
--- +++ @@ -0,0 +1,47 @@ +"""test_averaging_imputations.py""" +import unittest +import numpy as np +from impy.imputations import mean_imputation +from impy.imputations import mode_imputation +from impy.imputations import median_imputation +from impy.datasets import random_int + + +class TestAveraging(unittest.TestCase): + """ Tests for Averaging """ + def setUp(self): + self.data = random_int(missingness="complete") + + def test_mean_return_type(self): + """Mean Imputation Return Type""" + self.assertEqual(str(type(mean_imputation(self.data))), + "<class 'numpy.ndarray'>") + + def test_mode_return_type(self): + """Mode Imputation Return Type""" + self.assertEqual(str(type(mode_imputation(self.data))), + "<class 'numpy.ndarray'>") + + def test_median_return_type(self): + """Median Imputation Return Type""" + self.assertEqual(str(type(median_imputation(self.data))), + "<class 'numpy.ndarray'>") + + def test_mean_fill(self): + """ Mean Imputation Fill Complete Data(nothing should happen)""" + actual = mean_imputation(self.data) + self.assertTrue(np.array_equal(actual, self.data)) + + def test_mode_fill(self): + """ Mode Imputation Fill Complete Data(nothing should happen)""" + actual = mode_imputation(self.data) + self.assertTrue(np.array_equal(actual, self.data)) + + def test_median_fill(self): + """ Median Imputation Fill Complete Data(nothing should happen)""" + actual = median_imputation(self.data) + self.assertTrue(np.array_equal(actual, self.data)) + + +if __name__ == "__main__": + unittest.main()
17ae9e25663d029af11236584b4c759c895ae830
util/fileIngredients.py
util/fileIngredients.py
#!/usr/bin/env python # # This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this file, # You can obtain one at http://mozilla.org/MPL/2.0/. from __future__ import with_statement import re def fileContains(f, s, isRegex): if isRegex: return fileContainsRegex(f, re.compile(s, re.MULTILINE)) else: return fileContainsStr(f, s), s def fileContainsStr(f, s): found = False with open(f, 'rb') as g: for line in g: if line.find(s) != -1: print line.rstrip() found = True return found def fileContainsRegex(f, regex): # e.g. ~/fuzzing/lithium/lithium.py crashesat --timeout=30 # --regex '^#0\s*0x.* in\s*.*(?:\n|\r\n?)#1\s*' ./js --ion -n 735957.js # Note that putting "^" and "$" together is unlikely to work. matchedStr = '' found = False with open(f, 'rb') as g: foundRegex = regex.search(g.read()) if foundRegex: matchedStr = foundRegex.group() print matchedStr found = True return found, matchedStr
Improve and consolidate condition scripts of Lithium to support timeouts and regex via optparse. r=Jesse
Improve and consolidate condition scripts of Lithium to support timeouts and regex via optparse. r=Jesse
Python
mpl-2.0
nth10sd/funfuzz,nth10sd/funfuzz,MozillaSecurity/funfuzz,MozillaSecurity/funfuzz,nth10sd/funfuzz,MozillaSecurity/funfuzz
--- +++ @@ -0,0 +1,39 @@ +#!/usr/bin/env python +# +# This Source Code Form is subject to the terms of the Mozilla Public +# License, v. 2.0. If a copy of the MPL was not distributed with this file, +# You can obtain one at http://mozilla.org/MPL/2.0/. + +from __future__ import with_statement + +import re + +def fileContains(f, s, isRegex): + if isRegex: + return fileContainsRegex(f, re.compile(s, re.MULTILINE)) + else: + return fileContainsStr(f, s), s + + +def fileContainsStr(f, s): + found = False + with open(f, 'rb') as g: + for line in g: + if line.find(s) != -1: + print line.rstrip() + found = True + return found + +def fileContainsRegex(f, regex): + # e.g. ~/fuzzing/lithium/lithium.py crashesat --timeout=30 + # --regex '^#0\s*0x.* in\s*.*(?:\n|\r\n?)#1\s*' ./js --ion -n 735957.js + # Note that putting "^" and "$" together is unlikely to work. + matchedStr = '' + found = False + with open(f, 'rb') as g: + foundRegex = regex.search(g.read()) + if foundRegex: + matchedStr = foundRegex.group() + print matchedStr + found = True + return found, matchedStr
05aa314ac9b5d38bb7a30e30aced9b27b2797888
python/ql/test/experimental/dataflow/tainttracking/defaultAdditionalTaintStep/test_syntax.py
python/ql/test/experimental/dataflow/tainttracking/defaultAdditionalTaintStep/test_syntax.py
# Add taintlib to PATH so it can be imported during runtime without any hassle import sys; import os; sys.path.append(os.path.dirname(os.path.dirname((__file__)))) from taintlib import * # This has no runtime impact, but allows autocomplete to work from typing import TYPE_CHECKING if TYPE_CHECKING: from ..taintlib import * # Actual tests class Context: def __enter__(self): return TAINTED_STRING def __exit__(self, exc_type, exc, tb): pass def test_with(): ctx = Context() taint(ctx) with ctx as tainted: ensure_tainted(tainted) # $ tainted class Iter: def __iter__(self): return self def __next__(self): raise StopIteration def test_for(): iter = Iter() taint(iter) for tainted in iter: ensure_tainted(tainted) # $ tainted # Make tests runable test_with() test_for()
Add tests for non-async constructs
Python: Add tests for non-async constructs
Python
mit
github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql,github/codeql
--- +++ @@ -0,0 +1,45 @@ +# Add taintlib to PATH so it can be imported during runtime without any hassle +import sys; import os; sys.path.append(os.path.dirname(os.path.dirname((__file__)))) +from taintlib import * + +# This has no runtime impact, but allows autocomplete to work +from typing import TYPE_CHECKING +if TYPE_CHECKING: + from ..taintlib import * + + +# Actual tests + +class Context: + def __enter__(self): + return TAINTED_STRING + + def __exit__(self, exc_type, exc, tb): + pass + +def test_with(): + ctx = Context() + taint(ctx) + with ctx as tainted: + ensure_tainted(tainted) # $ tainted + + +class Iter: + def __iter__(self): + return self + + def __next__(self): + raise StopIteration + +def test_for(): + iter = Iter() + taint(iter) + for tainted in iter: + ensure_tainted(tainted) # $ tainted + + + +# Make tests runable + +test_with() +test_for()
c1fcf54b63de95c85a9505d83062d8b320b1cbdf
python/example_code/cloudfront/update_distribution_certificate.py
python/example_code/cloudfront/update_distribution_certificate.py
# Copyright 2010-2018 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"). You # may not use this file except in compliance with the License. A copy of # the License is located at # # http://aws.amazon.com/apache2.0/ # # or in the "license" file accompanying this file. This file is # distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF # ANY KIND, either express or implied. See the License for the specific # language governing permissions and limitations under the License. import boto3 import sys #support for python 2 and 3 input types def read(output): if sys.version_info[0] < 3: return(raw_input(output)) else: return(input(output)) # Create CloudFront client cf = boto3.client('cloudfront') # List distributions with the pagination interface print("\nAvailable CloudFront Distributions:\n") paginator = cf.get_paginator('list_distributions') for distributionlist in paginator.paginate(): for distribution in distributionlist['DistributionList']['Items']: print("Domain: " + distribution['DomainName']) print("Distribution Id: " + distribution['Id']) print("Certificate Source: " + distribution['ViewerCertificate']['CertificateSource']) if (distribution['ViewerCertificate']['CertificateSource'] == "acm"): print("Certificate ARN: " + distribution['ViewerCertificate']['Certificate']) print("") print('Enter the Distribution Id of the CloudFront Distribution who\'s ACM Certificate you would like to replace. ') distribution_id = read('Note that certificate source must be ACM - DistributionId: ') distribution_config_response=cf.get_distribution_config(Id=distribution_id) distribution_config=distribution_config_response['DistributionConfig'] distribution_etag=distribution_config_response['ETag'] if (distribution_config['ViewerCertificate']['CertificateSource'] != "acm"): print("\nThe DistributionId you have entered is not currently using an ACM Certificate, exiting...\n") exit() old_cert_arn=distribution_config['ViewerCertificate']['ACMCertificateArn'] new_cert_arn=read("Please enter the ARN of the new ACM Certificate you would like to attach to Distribution " + distribution_id + ": ") print("Replacing: " + old_cert_arn + "\nwith: " + new_cert_arn + "\n") distribution_config['ViewerCertificate']['ACMCertificateArn']=new_cert_arn distribution_config['ViewerCertificate']['Certificate']=new_cert_arn cf.update_distribution(DistributionConfig=distribution_config,Id=distribution_id,IfMatch=distribution_etag)
Add python cloudfront update_distribution example to replace ACM Certificate
Add python cloudfront update_distribution example to replace ACM Certificate Add python cloudfront update_distribution example to replace ACM Certificate
Python
apache-2.0
awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples,awsdocs/aws-doc-sdk-examples
--- +++ @@ -0,0 +1,59 @@ +# Copyright 2010-2018 Amazon.com, Inc. or its affiliates. All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"). You +# may not use this file except in compliance with the License. A copy of +# the License is located at +# +# http://aws.amazon.com/apache2.0/ +# +# or in the "license" file accompanying this file. This file is +# distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF +# ANY KIND, either express or implied. See the License for the specific +# language governing permissions and limitations under the License. + +import boto3 +import sys + +#support for python 2 and 3 input types +def read(output): + if sys.version_info[0] < 3: + return(raw_input(output)) + else: + return(input(output)) + +# Create CloudFront client +cf = boto3.client('cloudfront') + +# List distributions with the pagination interface +print("\nAvailable CloudFront Distributions:\n") +paginator = cf.get_paginator('list_distributions') +for distributionlist in paginator.paginate(): + for distribution in distributionlist['DistributionList']['Items']: + print("Domain: " + distribution['DomainName']) + print("Distribution Id: " + distribution['Id']) + print("Certificate Source: " + distribution['ViewerCertificate']['CertificateSource']) + if (distribution['ViewerCertificate']['CertificateSource'] == "acm"): + print("Certificate ARN: " + distribution['ViewerCertificate']['Certificate']) + print("") + +print('Enter the Distribution Id of the CloudFront Distribution who\'s ACM Certificate you would like to replace. ') +distribution_id = read('Note that certificate source must be ACM - DistributionId: ') + +distribution_config_response=cf.get_distribution_config(Id=distribution_id) +distribution_config=distribution_config_response['DistributionConfig'] +distribution_etag=distribution_config_response['ETag'] + +if (distribution_config['ViewerCertificate']['CertificateSource'] != "acm"): + print("\nThe DistributionId you have entered is not currently using an ACM Certificate, exiting...\n") + exit() + +old_cert_arn=distribution_config['ViewerCertificate']['ACMCertificateArn'] + +new_cert_arn=read("Please enter the ARN of the new ACM Certificate you would like to attach to Distribution " + distribution_id + ": ") + +print("Replacing: " + old_cert_arn + "\nwith: " + new_cert_arn + "\n") + +distribution_config['ViewerCertificate']['ACMCertificateArn']=new_cert_arn +distribution_config['ViewerCertificate']['Certificate']=new_cert_arn + +cf.update_distribution(DistributionConfig=distribution_config,Id=distribution_id,IfMatch=distribution_etag)
75a882bf38c88d73e38d13fbb8b1499ff4ae4ea6
scripts/remove_after_use/set_meetings_users_fullnames_to_guids.py
scripts/remove_after_use/set_meetings_users_fullnames_to_guids.py
import sys import logging import django from django.db import transaction django.setup() from osf.models import OSFUser logger = logging.getLogger(__name__) logging.basicConfig(level=logging.INFO) def main(): dry_run = '--dry' in sys.argv with transaction.atomic(): users = OSFUser.objects.filter(fullname__regex=r'^[A-Za-z0-9\.\+_-]+@[A-Za-z0-9\._-]+\.[a-zA-Z]*$', tags__name='osf4m') logger.info('{} users found added by OSF 4 Meetings with emails for fullnames'.format(users.count())) for user in users: user.fullname = user._id if not dry_run: user.save() if __name__ == '__main__': main()
Add migration for changing users added by OSF for meetings with emails for fullnames to their guid
Add migration for changing users added by OSF for meetings with emails for fullnames to their guid
Python
apache-2.0
felliott/osf.io,adlius/osf.io,felliott/osf.io,adlius/osf.io,saradbowman/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,Johnetordoff/osf.io,CenterForOpenScience/osf.io,mfraezz/osf.io,adlius/osf.io,felliott/osf.io,mattclark/osf.io,cslzchen/osf.io,CenterForOpenScience/osf.io,felliott/osf.io,pattisdr/osf.io,brianjgeiger/osf.io,aaxelb/osf.io,cslzchen/osf.io,baylee-d/osf.io,HalcyonChimera/osf.io,mfraezz/osf.io,aaxelb/osf.io,caseyrollins/osf.io,caseyrollins/osf.io,Johnetordoff/osf.io,aaxelb/osf.io,mfraezz/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,saradbowman/osf.io,caseyrollins/osf.io,cslzchen/osf.io,Johnetordoff/osf.io,HalcyonChimera/osf.io,mattclark/osf.io,CenterForOpenScience/osf.io,adlius/osf.io,aaxelb/osf.io,baylee-d/osf.io,cslzchen/osf.io,CenterForOpenScience/osf.io,brianjgeiger/osf.io,mattclark/osf.io,mfraezz/osf.io,baylee-d/osf.io,pattisdr/osf.io
--- +++ @@ -0,0 +1,24 @@ +import sys +import logging + +import django +from django.db import transaction +django.setup() + +from osf.models import OSFUser + +logger = logging.getLogger(__name__) +logging.basicConfig(level=logging.INFO) + +def main(): + dry_run = '--dry' in sys.argv + with transaction.atomic(): + users = OSFUser.objects.filter(fullname__regex=r'^[A-Za-z0-9\.\+_-]+@[A-Za-z0-9\._-]+\.[a-zA-Z]*$', tags__name='osf4m') + logger.info('{} users found added by OSF 4 Meetings with emails for fullnames'.format(users.count())) + for user in users: + user.fullname = user._id + if not dry_run: + user.save() + +if __name__ == '__main__': + main()
a6b35a9a94b2e4b32c2236258812b44e81184515
corehq/apps/users/management/commands/fix_location_user_data.py
corehq/apps/users/management/commands/fix_location_user_data.py
from corehq.apps.locations.models import Location from corehq.apps.users.models import CommCareUser from dimagi.utils.couch.database import iter_docs from django.core.management.base import BaseCommand, CommandError class Command(BaseCommand): args = "domain" help = "Fix location user data for mobile workers." def process_user(self, user): if user.location_id: user.set_location(Location.get(user.location_id)) else: user.unset_location() def handle(self, *args, **options): if len(args) == 0: raise CommandError("Usage: python manage.py fix_location_user_data %s" % self.args) domain = args[0] ids = ( CommCareUser.ids_by_domain(domain, is_active=True) + CommCareUser.ids_by_domain(domain, is_active=False) ) for doc in iter_docs(CommCareUser.get_db(), ids): user = CommCareUser.wrap(doc) try: self.process_user(user) except Exception as e: print "Error processing user %s: %s" % (user._id, e)
Add management command for resyncing mobile worker location user data
Add management command for resyncing mobile worker location user data
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq,qedsoftware/commcare-hq,qedsoftware/commcare-hq,dimagi/commcare-hq
--- +++ @@ -0,0 +1,31 @@ +from corehq.apps.locations.models import Location +from corehq.apps.users.models import CommCareUser +from dimagi.utils.couch.database import iter_docs +from django.core.management.base import BaseCommand, CommandError + + +class Command(BaseCommand): + args = "domain" + help = "Fix location user data for mobile workers." + + def process_user(self, user): + if user.location_id: + user.set_location(Location.get(user.location_id)) + else: + user.unset_location() + + def handle(self, *args, **options): + if len(args) == 0: + raise CommandError("Usage: python manage.py fix_location_user_data %s" % self.args) + + domain = args[0] + ids = ( + CommCareUser.ids_by_domain(domain, is_active=True) + + CommCareUser.ids_by_domain(domain, is_active=False) + ) + for doc in iter_docs(CommCareUser.get_db(), ids): + user = CommCareUser.wrap(doc) + try: + self.process_user(user) + except Exception as e: + print "Error processing user %s: %s" % (user._id, e)
415717bddb00ca650bef61a5c6054a7b47575b56
jaspyx/tests/visitor/test_break.py
jaspyx/tests/visitor/test_break.py
import ast from jaspyx.ast_util import ast_store, ast_load from jaspyx.tests.visitor.v8_helper import V8Helper class TestBreak(V8Helper): def test_break(self): assert self.run( [ ast.Assign( [ast_store('i')], ast.Num(0), ), ast.While( ast.Compare( ast_load('i'), [ast.Lt()], [ast.Num(10)] ), [ ast.Break(), ], [] ) ], 'i', int ) == 0
Implement unit test for break.
Implement unit test for break.
Python
mit
iksteen/jaspyx,ztane/jaspyx
--- +++ @@ -0,0 +1,28 @@ +import ast +from jaspyx.ast_util import ast_store, ast_load +from jaspyx.tests.visitor.v8_helper import V8Helper + + +class TestBreak(V8Helper): + def test_break(self): + assert self.run( + [ + ast.Assign( + [ast_store('i')], + ast.Num(0), + ), + ast.While( + ast.Compare( + ast_load('i'), + [ast.Lt()], + [ast.Num(10)] + ), + [ + ast.Break(), + ], + [] + ) + ], + 'i', + int + ) == 0
495da73f305a2a0e79a28d251b5b93caea06656d
mediagenerator/filters/uglifier.py
mediagenerator/filters/uglifier.py
from django.conf import settings from django.utils.encoding import smart_str from mediagenerator.generators.bundles.base import Filter class Uglifier(Filter): def __init__(self, **kwargs): super(Uglifier, self).__init__(**kwargs) assert self.filetype == 'js', ( 'Uglifier only supports compilation to js. ' 'The parent filter expects "%s".' % self.filetype) def get_output(self, variation): # We import this here, so App Engine Helper users don't get import # errors. from subprocess import Popen, PIPE for input in self.get_input(variation): args = ['uglifyjs'] try: args = args + settings.UGLIFIER_OPTIONS except AttributeError: pass try: cmd = Popen(args, stdin=PIPE, stdout=PIPE, stderr=PIPE, universal_newlines=True) output, error = cmd.communicate(smart_str(input)) assert cmd.wait() == 0, 'Command returned bad result:\n%s' % error yield output.decode('utf-8') except Exception, e: raise ValueError("Failed to run UglifyJs. " "Please make sure you have Node.js and UglifyJS installed " "and that it's in your PATH.\n" "Error was: %s" % e)
Add UglifyJS as a filter.
Add UglifyJS as a filter.
Python
bsd-3-clause
Carlangueitor/django-mediagenerator,adieu/django-mediagenerator,Carlangueitor/django-mediagenerator,adieu/django-mediagenerator
--- +++ @@ -0,0 +1,33 @@ +from django.conf import settings +from django.utils.encoding import smart_str +from mediagenerator.generators.bundles.base import Filter + +class Uglifier(Filter): + def __init__(self, **kwargs): + super(Uglifier, self).__init__(**kwargs) + assert self.filetype == 'js', ( + 'Uglifier only supports compilation to js. ' + 'The parent filter expects "%s".' % self.filetype) + + def get_output(self, variation): + # We import this here, so App Engine Helper users don't get import + # errors. + from subprocess import Popen, PIPE + for input in self.get_input(variation): + args = ['uglifyjs'] + try: + args = args + settings.UGLIFIER_OPTIONS + except AttributeError: + pass + try: + cmd = Popen(args, + stdin=PIPE, stdout=PIPE, stderr=PIPE, + universal_newlines=True) + output, error = cmd.communicate(smart_str(input)) + assert cmd.wait() == 0, 'Command returned bad result:\n%s' % error + yield output.decode('utf-8') + except Exception, e: + raise ValueError("Failed to run UglifyJs. " + "Please make sure you have Node.js and UglifyJS installed " + "and that it's in your PATH.\n" + "Error was: %s" % e)
96ed06f1f3dab3aa9d0f8150c41a5c1b943a86b0
frappe/tests/test_config.py
frappe/tests/test_config.py
# Copyright (c) 2022, Frappe Technologies Pvt. Ltd. and Contributors # License: MIT. See LICENSE import unittest import frappe from frappe.config import get_modules_from_all_apps_for_user class TestConfig(unittest.TestCase): def test_get_modules(self): frappe_modules = frappe.get_all("Module Def", filters={"app_name": "frappe"}, pluck="name") all_modules_data = get_modules_from_all_apps_for_user() first_module_entry = all_modules_data[0] all_modules = [x["module_name"] for x in all_modules_data] self.assertIn("links", first_module_entry) self.assertIsInstance(all_modules_data, list) self.assertFalse([x for x in frappe_modules if x not in all_modules])
Add test for config module
test: Add test for config module
Python
mit
StrellaGroup/frappe,frappe/frappe,StrellaGroup/frappe,frappe/frappe,StrellaGroup/frappe,frappe/frappe
--- +++ @@ -0,0 +1,17 @@ +# Copyright (c) 2022, Frappe Technologies Pvt. Ltd. and Contributors +# License: MIT. See LICENSE +import unittest + +import frappe +from frappe.config import get_modules_from_all_apps_for_user + + +class TestConfig(unittest.TestCase): + def test_get_modules(self): + frappe_modules = frappe.get_all("Module Def", filters={"app_name": "frappe"}, pluck="name") + all_modules_data = get_modules_from_all_apps_for_user() + first_module_entry = all_modules_data[0] + all_modules = [x["module_name"] for x in all_modules_data] + self.assertIn("links", first_module_entry) + self.assertIsInstance(all_modules_data, list) + self.assertFalse([x for x in frappe_modules if x not in all_modules])
f1ccab2168dea1b0827f4ca929f0036e84170a76
go/base/tests/test_views.py
go/base/tests/test_views.py
"""Test for go.base.utils.""" from mock import patch, Mock from django.core.urlresolvers import reverse from go.base.tests.utils import VumiGoDjangoTestCase class BaseViewsTestCase(VumiGoDjangoTestCase): def cross_domain_xhr(self, url): return self.client.post(reverse('cross_domain_xhr'), {'url': url}) @patch('requests.get') def test_cross_domain_xhr(self, mocked_get): mocked_get.return_value = Mock(content='foo', status_code=200) response = self.cross_domain_xhr('http://domain.com') [call] = mocked_get.call_args_list args, kwargs = call self.assertEqual(args, ('http://domain.com',)) self.assertEqual(kwargs, {'auth': None}) self.assertTrue(mocked_get.called) self.assertEqual(response.content, 'foo') self.assertEqual(response.status_code, 200) @patch('requests.get') def test_basic_auth_cross_domain_xhr(self, mocked_get): mocked_get.return_value = Mock(content='foo', status_code=200) response = self.cross_domain_xhr('http://username:password@domain.com') [call] = mocked_get.call_args_list args, kwargs = call self.assertEqual(args, ('http://domain.com',)) self.assertEqual(kwargs, {'auth': ('username', 'password')}) self.assertTrue(mocked_get.called) self.assertEqual(response.content, 'foo') self.assertEqual(response.status_code, 200) @patch('requests.get') def test_basic_auth_cross_domain_xhr_with_https_and_port(self, mocked_get): mocked_get.return_value = Mock(content='foo', status_code=200) response = self.cross_domain_xhr( 'https://username:password@domain.com:443/foo') [call] = mocked_get.call_args_list args, kwargs = call self.assertEqual(args, ('https://domain.com:443/foo',)) self.assertEqual(kwargs, {'auth': ('username', 'password')}) self.assertTrue(mocked_get.called) self.assertEqual(response.content, 'foo') self.assertEqual(response.status_code, 200)
Add tests for cross domain xhr view
Add tests for cross domain xhr view
Python
bsd-3-clause
praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go
--- +++ @@ -0,0 +1,48 @@ +"""Test for go.base.utils.""" + +from mock import patch, Mock +from django.core.urlresolvers import reverse + +from go.base.tests.utils import VumiGoDjangoTestCase + + +class BaseViewsTestCase(VumiGoDjangoTestCase): + def cross_domain_xhr(self, url): + return self.client.post(reverse('cross_domain_xhr'), {'url': url}) + + @patch('requests.get') + def test_cross_domain_xhr(self, mocked_get): + mocked_get.return_value = Mock(content='foo', status_code=200) + response = self.cross_domain_xhr('http://domain.com') + [call] = mocked_get.call_args_list + args, kwargs = call + self.assertEqual(args, ('http://domain.com',)) + self.assertEqual(kwargs, {'auth': None}) + self.assertTrue(mocked_get.called) + self.assertEqual(response.content, 'foo') + self.assertEqual(response.status_code, 200) + + @patch('requests.get') + def test_basic_auth_cross_domain_xhr(self, mocked_get): + mocked_get.return_value = Mock(content='foo', status_code=200) + response = self.cross_domain_xhr('http://username:password@domain.com') + [call] = mocked_get.call_args_list + args, kwargs = call + self.assertEqual(args, ('http://domain.com',)) + self.assertEqual(kwargs, {'auth': ('username', 'password')}) + self.assertTrue(mocked_get.called) + self.assertEqual(response.content, 'foo') + self.assertEqual(response.status_code, 200) + + @patch('requests.get') + def test_basic_auth_cross_domain_xhr_with_https_and_port(self, mocked_get): + mocked_get.return_value = Mock(content='foo', status_code=200) + response = self.cross_domain_xhr( + 'https://username:password@domain.com:443/foo') + [call] = mocked_get.call_args_list + args, kwargs = call + self.assertEqual(args, ('https://domain.com:443/foo',)) + self.assertEqual(kwargs, {'auth': ('username', 'password')}) + self.assertTrue(mocked_get.called) + self.assertEqual(response.content, 'foo') + self.assertEqual(response.status_code, 200)
5578d11f45e9c41ab9c4311f2bed48b9c24d9bf5
tests/grammar_term-nonterm_test/NonterminalHaveTest.py
tests/grammar_term-nonterm_test/NonterminalHaveTest.py
#!/usr/bin/env python """ :Author Patrik Valkovic :Created 23.06.2017 16:39 :Licence GNUv3 Part of grammpy """
Create file for Nonterminal have method
Create file for Nonterminal have method
Python
mit
PatrikValkovic/grammpy
--- +++ @@ -0,0 +1,8 @@ +#!/usr/bin/env python +""" +:Author Patrik Valkovic +:Created 23.06.2017 16:39 +:Licence GNUv3 +Part of grammpy + +"""
913c9a10b2eb3b3d9de108a82a3251b2c0de0e10
cybox/test/objects/hostname_test.py
cybox/test/objects/hostname_test.py
# Copyright (c) 2014, The MITRE Corporation. All rights reserved. # See LICENSE.txt for complete terms. import unittest from cybox.objects.hostname_object import Hostname from cybox.test.objects import ObjectTestCase class TestHostname(ObjectTestCase, unittest.TestCase): object_type = "HostnameObjectType" klass = Hostname _full_dict = { 'is_domain_name': True, 'hostname_value': "www.example.com", 'naming_system': ["DNS", "NETBIOS"], 'xsi:type': object_type, } if __name__ == "__main__": unittest.main()
Add test for Hostname object
Add test for Hostname object
Python
bsd-3-clause
CybOXProject/python-cybox
--- +++ @@ -0,0 +1,22 @@ +# Copyright (c) 2014, The MITRE Corporation. All rights reserved. +# See LICENSE.txt for complete terms. + +import unittest + +from cybox.objects.hostname_object import Hostname +from cybox.test.objects import ObjectTestCase + + +class TestHostname(ObjectTestCase, unittest.TestCase): + object_type = "HostnameObjectType" + klass = Hostname + + _full_dict = { + 'is_domain_name': True, + 'hostname_value': "www.example.com", + 'naming_system': ["DNS", "NETBIOS"], + 'xsi:type': object_type, + } + +if __name__ == "__main__": + unittest.main()
514aca20c6f076a86819d7180f36c3b2e8bcc33b
tests/integration_tests/test_tensorflow_integration.py
tests/integration_tests/test_tensorflow_integration.py
from __future__ import print_function import os import tempfile import pytest import keras from keras import layers from keras.utils.test_utils import get_test_data from keras.utils.test_utils import keras_test @pytest.mark.skipif(keras.backend.backend() != 'tensorflow', reason='Requires TF backend') @keras_test def test_tf_optimizer(): import tensorflow as tf num_hidden = 10 output_dim = 2 input_dim = 10 target = 0.8 optimizer = tf.train.AdadeltaOptimizer( learning_rate=1., rho=0.95, epsilon=1e-08) (x_train, y_train), (x_test, y_test) = get_test_data( num_train=1000, num_test=200, input_shape=(input_dim,), classification=True, num_classes=output_dim) model = keras.Sequential() model.add(layers.Dense(num_hidden, activation='relu', input_shape=(input_dim,))) model.add(layers.Dense(output_dim, activation='softmax')) model.compile(loss='sparse_categorical_crossentropy', optimizer=optimizer, metrics=['accuracy']) history = model.fit(x_train, y_train, epochs=8, batch_size=16, validation_data=(x_test, y_test), verbose=2) assert history.history['val_acc'][-1] >= target # Test saving. _, fname = tempfile.mkstemp('.h5') model.save(fname) model = keras.models.load_model(fname) assert len(model.weights) == 4 os.remove(fname) if __name__ == '__main__': pytest.main([__file__])
Add integration test checking compatibility of Keras models with TF optimizers.
Add integration test checking compatibility of Keras models with TF optimizers.
Python
apache-2.0
keras-team/keras,keras-team/keras
--- +++ @@ -0,0 +1,51 @@ +from __future__ import print_function + +import os +import tempfile +import pytest +import keras +from keras import layers +from keras.utils.test_utils import get_test_data +from keras.utils.test_utils import keras_test + + +@pytest.mark.skipif(keras.backend.backend() != 'tensorflow', reason='Requires TF backend') +@keras_test +def test_tf_optimizer(): + import tensorflow as tf + + num_hidden = 10 + output_dim = 2 + input_dim = 10 + target = 0.8 + optimizer = tf.train.AdadeltaOptimizer( + learning_rate=1., rho=0.95, epsilon=1e-08) + + (x_train, y_train), (x_test, y_test) = get_test_data( + num_train=1000, num_test=200, + input_shape=(input_dim,), + classification=True, num_classes=output_dim) + + model = keras.Sequential() + model.add(layers.Dense(num_hidden, + activation='relu', + input_shape=(input_dim,))) + model.add(layers.Dense(output_dim, activation='softmax')) + + model.compile(loss='sparse_categorical_crossentropy', + optimizer=optimizer, + metrics=['accuracy']) + history = model.fit(x_train, y_train, epochs=8, batch_size=16, + validation_data=(x_test, y_test), verbose=2) + assert history.history['val_acc'][-1] >= target + + # Test saving. + _, fname = tempfile.mkstemp('.h5') + model.save(fname) + model = keras.models.load_model(fname) + assert len(model.weights) == 4 + os.remove(fname) + + +if __name__ == '__main__': + pytest.main([__file__])
24d742e444c84df99629d8a6aff7ca7e6c90f995
scheduler/misc/detect_stuck_active_invs.py
scheduler/misc/detect_stuck_active_invs.py
#!/usr/bin/env python # Copyright 2018 The LUCI Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Finds jobs with old entries (>1d) in ActiveInvocations list. Usage: prpc login ./detect_stuck_active_invs.py luci-scheduler-dev.appspot.com Requires the caller to be in 'administrators' group. """ import json import subprocess import sys import time def prpc(host, method, body): p = subprocess.Popen( ['prpc', 'call', host, method], stdin=subprocess.PIPE, stdout=subprocess.PIPE) out, _ = p.communicate(json.dumps(body)) if p.returncode: raise Exception('Call to %s failed' % method) return json.loads(out) def check_job(host, job_ref): print 'Checking %s/%s' % (job_ref['project'], job_ref['job']) state = prpc(host, 'internal.admin.Admin.GetDebugJobState', job_ref) active_invs = state.get('activeInvocations', []) if not active_invs: print ' No active invocations' return [] stuck = [] for inv_id in active_invs: print ' ...checking %s' % inv_id inv = prpc(host, 'scheduler.Scheduler.GetInvocation', { 'jobRef': job_ref, 'invocationId': inv_id, }) started = time.time() - int(inv['startedTs']) / 1000000.0 if started > 24 * 3600: print ' it is stuck!' stuck.append((job_ref, inv_id)) return stuck def main(): if len(sys.argv) != 2: print >> sys.stderr, 'Usage: %s <host>' % sys.argv[0] return 1 host = sys.argv[1] stuck = [] for job in prpc(host, 'scheduler.Scheduler.GetJobs', {})['jobs']: stuck.extend(check_job(host, job['jobRef'])) if not stuck: print 'No invocations are stuck' return print print 'All stuck invocations: ' for job_ref, inv_id in stuck: print '%s/%s %s' % (job_ref['project'], job_ref['job'], inv_id) return 0 if __name__ == '__main__': sys.exit(main())
Add adhoc script to detect jobs with stuck ActiveInvocations list.
[scheduler] Add adhoc script to detect jobs with stuck ActiveInvocations list. R=a30c74fa30536fe7ea81ed6dec202e35e149e1fd@chromium.org BUG=852142 Change-Id: Idae7f05c5045a72ff85db8587f8bd74c0b80fb06 Reviewed-on: https://chromium-review.googlesource.com/1098463 Reviewed-by: Andrii Shyshkalov <a30c74fa30536fe7ea81ed6dec202e35e149e1fd@chromium.org> Commit-Queue: Vadim Shtayura <9f116ddb1b24f6fc1916a676eb17161b6c07dfc1@chromium.org>
Python
apache-2.0
luci/luci-go,luci/luci-go,luci/luci-go,luci/luci-go,luci/luci-go,luci/luci-go
--- +++ @@ -0,0 +1,88 @@ +#!/usr/bin/env python +# Copyright 2018 The LUCI Authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""Finds jobs with old entries (>1d) in ActiveInvocations list. + +Usage: + prpc login + ./detect_stuck_active_invs.py luci-scheduler-dev.appspot.com + +Requires the caller to be in 'administrators' group. +""" + +import json +import subprocess +import sys +import time + + +def prpc(host, method, body): + p = subprocess.Popen( + ['prpc', 'call', host, method], + stdin=subprocess.PIPE, + stdout=subprocess.PIPE) + out, _ = p.communicate(json.dumps(body)) + if p.returncode: + raise Exception('Call to %s failed' % method) + return json.loads(out) + + +def check_job(host, job_ref): + print 'Checking %s/%s' % (job_ref['project'], job_ref['job']) + + state = prpc(host, 'internal.admin.Admin.GetDebugJobState', job_ref) + active_invs = state.get('activeInvocations', []) + if not active_invs: + print ' No active invocations' + return [] + + stuck = [] + for inv_id in active_invs: + print ' ...checking %s' % inv_id + inv = prpc(host, 'scheduler.Scheduler.GetInvocation', { + 'jobRef': job_ref, + 'invocationId': inv_id, + }) + started = time.time() - int(inv['startedTs']) / 1000000.0 + if started > 24 * 3600: + print ' it is stuck!' + stuck.append((job_ref, inv_id)) + return stuck + + +def main(): + if len(sys.argv) != 2: + print >> sys.stderr, 'Usage: %s <host>' % sys.argv[0] + return 1 + host = sys.argv[1] + + stuck = [] + for job in prpc(host, 'scheduler.Scheduler.GetJobs', {})['jobs']: + stuck.extend(check_job(host, job['jobRef'])) + + if not stuck: + print 'No invocations are stuck' + return + + print + print 'All stuck invocations: ' + for job_ref, inv_id in stuck: + print '%s/%s %s' % (job_ref['project'], job_ref['job'], inv_id) + + return 0 + + +if __name__ == '__main__': + sys.exit(main())
cf78037980a9345c12b1e2562bc4eda63cea95b3
test/functionalities/backticks/TestBackticksWithoutATarget.py
test/functionalities/backticks/TestBackticksWithoutATarget.py
""" Test that backticks without a target should work (not infinite looping). """ import os, time import unittest2 import lldb from lldbtest import * class BackticksWithNoTargetTestCase(TestBase): mydir = "functionalities/backticks" def test_backticks_no_target(self): """A simple test of backticks without a target.""" self.expect("print `1+2-3`", substrs = [' = 0']) if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
Add a simple regression test to go with r143260. CommandInterpreter::PreprocessCommand() should not infinite loop when a target has not been specified yet.
Add a simple regression test to go with r143260. CommandInterpreter::PreprocessCommand() should not infinite loop when a target has not been specified yet. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@143274 91177308-0d34-0410-b5e6-96231b3b80d8
Python
apache-2.0
apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb
--- +++ @@ -0,0 +1,23 @@ +""" +Test that backticks without a target should work (not infinite looping). +""" + +import os, time +import unittest2 +import lldb +from lldbtest import * + +class BackticksWithNoTargetTestCase(TestBase): + + mydir = "functionalities/backticks" + + def test_backticks_no_target(self): + """A simple test of backticks without a target.""" + self.expect("print `1+2-3`", + substrs = [' = 0']) + +if __name__ == '__main__': + import atexit + lldb.SBDebugger.Initialize() + atexit.register(lambda: lldb.SBDebugger.Terminate()) + unittest2.main()
ff2fb40e961fe7b0c3f6dd6e91d8fb79a865a631
src/ggrc_basic_permissions/migrations/versions/20131108224846_37b63b122038_hide_the_auditorread.py
src/ggrc_basic_permissions/migrations/versions/20131108224846_37b63b122038_hide_the_auditorread.py
"""Hide the AuditorReader role by changing its scope. Revision ID: 37b63b122038 Revises: 1ff082d26157 Create Date: 2013-11-08 22:48:46.956836 """ # revision identifiers, used by Alembic. revision = '37b63b122038' down_revision = '1ff082d26157' import sqlalchemy as sa from alembic import op from datetime import datetime from sqlalchemy.sql import table, column roles_table = table('roles', column('id', sa.Integer), column('name', sa.String), column('permissions_json', sa.Text), column('description', sa.Text), column('modified_by_id', sa.Integer), column('created_at', sa.DateTime), column('updated_at', sa.DateTime), column('context_id', sa.Integer), column('scope', sa.String), ) def upgrade(): op.execute( roles_table.update()\ .where(roles_table.c.name == 'AuditorReader')\ .values({'scope': 'System Implied'}) ) def downgrade(): op.execute( roles_table.update()\ .where(roles_table.c.name == 'AuditorReader')\ .values({'scope': 'System'}) )
Hide the AuditorReader role in the system role assignment modal by placing it in an "implied" scope.
Hide the AuditorReader role in the system role assignment modal by placing it in an "implied" scope.
Python
apache-2.0
josthkko/ggrc-core,uskudnik/ggrc-core,selahssea/ggrc-core,prasannav7/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,kr41/ggrc-core,kr41/ggrc-core,vladan-m/ggrc-core,NejcZupec/ggrc-core,edofic/ggrc-core,hasanalom/ggrc-core,hasanalom/ggrc-core,andrei-karalionak/ggrc-core,selahssea/ggrc-core,josthkko/ggrc-core,andrei-karalionak/ggrc-core,plamut/ggrc-core,AleksNeStu/ggrc-core,j0gurt/ggrc-core,josthkko/ggrc-core,uskudnik/ggrc-core,edofic/ggrc-core,edofic/ggrc-core,VinnieJohns/ggrc-core,j0gurt/ggrc-core,andrei-karalionak/ggrc-core,jmakov/ggrc-core,josthkko/ggrc-core,kr41/ggrc-core,selahssea/ggrc-core,jmakov/ggrc-core,NejcZupec/ggrc-core,prasannav7/ggrc-core,hyperNURb/ggrc-core,jmakov/ggrc-core,vladan-m/ggrc-core,vladan-m/ggrc-core,vladan-m/ggrc-core,hyperNURb/ggrc-core,jmakov/ggrc-core,plamut/ggrc-core,uskudnik/ggrc-core,hyperNURb/ggrc-core,AleksNeStu/ggrc-core,kr41/ggrc-core,prasannav7/ggrc-core,VinnieJohns/ggrc-core,vladan-m/ggrc-core,VinnieJohns/ggrc-core,selahssea/ggrc-core,hyperNURb/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,plamut/ggrc-core,hyperNURb/ggrc-core,uskudnik/ggrc-core,NejcZupec/ggrc-core,AleksNeStu/ggrc-core,andrei-karalionak/ggrc-core,hasanalom/ggrc-core,hasanalom/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core,jmakov/ggrc-core,edofic/ggrc-core
--- +++ @@ -0,0 +1,43 @@ + +"""Hide the AuditorReader role by changing its scope. + +Revision ID: 37b63b122038 +Revises: 1ff082d26157 +Create Date: 2013-11-08 22:48:46.956836 + +""" + +# revision identifiers, used by Alembic. +revision = '37b63b122038' +down_revision = '1ff082d26157' + +import sqlalchemy as sa +from alembic import op +from datetime import datetime +from sqlalchemy.sql import table, column + +roles_table = table('roles', + column('id', sa.Integer), + column('name', sa.String), + column('permissions_json', sa.Text), + column('description', sa.Text), + column('modified_by_id', sa.Integer), + column('created_at', sa.DateTime), + column('updated_at', sa.DateTime), + column('context_id', sa.Integer), + column('scope', sa.String), + ) + +def upgrade(): + op.execute( + roles_table.update()\ + .where(roles_table.c.name == 'AuditorReader')\ + .values({'scope': 'System Implied'}) + ) + +def downgrade(): + op.execute( + roles_table.update()\ + .where(roles_table.c.name == 'AuditorReader')\ + .values({'scope': 'System'}) + )
6e535a2d597f172d9342fb8a547335890c474b49
src/config-sample.py
src/config-sample.py
FLASK_SECRET_KEY = 'Enter a Flask Secret Key' # OAuth Credentials. You can find them on # https://www.yelp.com/developers/v3/manage_app YELP_CLIENT_ID = 'Enter Yelp Client ID' YELP_CLIENT_SECRET = 'Enter Yelp Client Secret'
Add a sample config file
Add a sample config file
Python
mit
byanofsky/playa-vista-neighborhood,byanofsky/playa-vista-neighborhood,byanofsky/playa-vista-neighborhood
--- +++ @@ -0,0 +1,6 @@ +FLASK_SECRET_KEY = 'Enter a Flask Secret Key' + +# OAuth Credentials. You can find them on +# https://www.yelp.com/developers/v3/manage_app +YELP_CLIENT_ID = 'Enter Yelp Client ID' +YELP_CLIENT_SECRET = 'Enter Yelp Client Secret'
22298d91fff788c37395cdad9245b3e7ed20cfdf
python/opencv/opencv_2/images/display_image_with_matplotlib.py
python/opencv/opencv_2/images/display_image_with_matplotlib.py
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (c) 2015 Jérémie DECOCK (http://www.jdhp.org) """ OpenCV - Display image: display an image given in arguments Required: opencv library (Debian: aptitude install python-opencv) See: https://opencv-python-tutroals.readthedocs.org/en/latest/py_tutorials/py_gui/py_image_display/py_image_display.html#display-an-image """ from __future__ import print_function import cv2 as cv import argparse from matplotlib import pyplot as plt def main(): # Parse the programm options (get the path of the image file to display) parser = argparse.ArgumentParser(description='An opencv snippet.') parser.add_argument("--infile", "-i", help="The picture file to display", required=True, metavar="FILE") args = parser.parse_args() infile_str = args.infile # OpenCV # imread_flags is a flag which specifies the way image should be read: # - cv.IMREAD_COLOR loads a color image. Any transparency of image will be neglected. It is the default flag. # - cv.IMREAD_GRAYSCALE loads image in grayscale mode # - cv.IMREAD_UNCHANGED loads image as such including alpha channel imread_flags = cv.IMREAD_GRAYSCALE img_np = cv.imread(infile_str, imread_flags) plt.imshow(img_np, cmap='gray', interpolation='none') # Display the image "img_np" with matplotlib plt.xticks([]) # to hide tick values on X axis plt.yticks([]) # to hide tick values on Y axis plt.show() if __name__ == '__main__': main()
Add a snippet (Python OpenCV).
Add a snippet (Python OpenCV).
Python
mit
jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets,jeremiedecock/snippets
--- +++ @@ -0,0 +1,45 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +# Copyright (c) 2015 Jérémie DECOCK (http://www.jdhp.org) + +""" +OpenCV - Display image: display an image given in arguments + +Required: opencv library (Debian: aptitude install python-opencv) + +See: https://opencv-python-tutroals.readthedocs.org/en/latest/py_tutorials/py_gui/py_image_display/py_image_display.html#display-an-image +""" + +from __future__ import print_function + +import cv2 as cv +import argparse +from matplotlib import pyplot as plt + +def main(): + + # Parse the programm options (get the path of the image file to display) + + parser = argparse.ArgumentParser(description='An opencv snippet.') + parser.add_argument("--infile", "-i", help="The picture file to display", required=True, metavar="FILE") + args = parser.parse_args() + + infile_str = args.infile + + # OpenCV + + # imread_flags is a flag which specifies the way image should be read: + # - cv.IMREAD_COLOR loads a color image. Any transparency of image will be neglected. It is the default flag. + # - cv.IMREAD_GRAYSCALE loads image in grayscale mode + # - cv.IMREAD_UNCHANGED loads image as such including alpha channel + imread_flags = cv.IMREAD_GRAYSCALE + img_np = cv.imread(infile_str, imread_flags) + + plt.imshow(img_np, cmap='gray', interpolation='none') # Display the image "img_np" with matplotlib + plt.xticks([]) # to hide tick values on X axis + plt.yticks([]) # to hide tick values on Y axis + plt.show() + +if __name__ == '__main__': + main()
ba9e4c6b003cc002e5bc7216da960e47f9fe5424
copper_imidazole_csv_allnitrogen.py
copper_imidazole_csv_allnitrogen.py
#!/usr/bin/env python2 import orca_parser from copper_imidazole_analysis import CopperImidazoleAnalysis import argparse import csv cia = CopperImidazoleAnalysis() parser = argparse.ArgumentParser(description="Given pathnames of ORCA output files, make a dump of all nitrogen parameters to a CSV file.") parser.add_argument("--csvname", dest="csvname", metavar="<CSV output root name>", type=str, default="nitrogen.csv", help="optional name for the CSV output file") parser.add_argument(dest="namelist", metavar="<ORCA filename>", nargs="+", type=str, default=None, help="ORCA output files") args = parser.parse_args() namelist = args.namelist with open(args.csvname, 'wb') as csvfile: csvwriter = csv.writer(csvfile, delimiter=',') for name in namelist: csvwriter.writerow([name]) csvwriter.writerow(["g-tensor", "id_copper", "A_copper (MHz)", "euler_copper (deg.)", "NQCC_copper (MHz)", "eta_copper"]) orcafile = orca_parser.ORCAOutputParser(name) gtensor, giso = orcafile.return_gtensor() id_copper = cia.copper_id(orcafile) atensor_copper = cia.hyperfine(orcafile, id_copper) euler_copper = cia.euler(orcafile, id_copper) nqi_copper, nqcc_copper, eta_copper = cia.nqi(orcafile, id_copper) csvwriter.writerow([gtensor, id_copper, atensor_copper, euler_copper, nqcc_copper, eta_copper]) csvwriter.writerow(["", "id_nitrogen", "A_nitrogen (MHz)", "euler_nitrogen (deg.)", "NQCC_nitrogen (MHz)", "eta_nitrogen", "Cu_N_distance (Angstroms)"]) nitrogen_list = orcafile.find_element("N") for id_nitrogen in nitrogen_list: atensor_nitrogen = cia.hyperfine(orcafile, id_nitrogen) euler_nitrogen = cia.euler(orcafile, id_nitrogen) nqi_nitrogen, nqcc_nitrogen, eta_nitrogen = cia.nqi(orcafile, id_nitrogen) cu_n_dist = orcafile.pair_distance(id_copper, id_nitrogen) csvwriter.writerow(["", id_nitrogen, atensor_nitrogen, euler_nitrogen, nqcc_nitrogen, eta_nitrogen, cu_n_dist])
Print information about all nitrogens.
Print information about all nitrogens.
Python
mpl-2.0
berquist/orcaparse
--- +++ @@ -0,0 +1,66 @@ +#!/usr/bin/env python2 + +import orca_parser +from copper_imidazole_analysis import CopperImidazoleAnalysis +import argparse +import csv + +cia = CopperImidazoleAnalysis() + +parser = argparse.ArgumentParser(description="Given pathnames of ORCA output files, make a dump of all nitrogen parameters to a CSV file.") + +parser.add_argument("--csvname", dest="csvname", metavar="<CSV output root name>", type=str, default="nitrogen.csv", help="optional name for the CSV output file") +parser.add_argument(dest="namelist", metavar="<ORCA filename>", nargs="+", type=str, default=None, help="ORCA output files") + +args = parser.parse_args() +namelist = args.namelist + +with open(args.csvname, 'wb') as csvfile: + csvwriter = csv.writer(csvfile, delimiter=',') + + for name in namelist: + csvwriter.writerow([name]) + csvwriter.writerow(["g-tensor", + "id_copper", + "A_copper (MHz)", + "euler_copper (deg.)", + "NQCC_copper (MHz)", + "eta_copper"]) + + orcafile = orca_parser.ORCAOutputParser(name) + gtensor, giso = orcafile.return_gtensor() + id_copper = cia.copper_id(orcafile) + atensor_copper = cia.hyperfine(orcafile, id_copper) + euler_copper = cia.euler(orcafile, id_copper) + nqi_copper, nqcc_copper, eta_copper = cia.nqi(orcafile, id_copper) + + csvwriter.writerow([gtensor, + id_copper, + atensor_copper, + euler_copper, + nqcc_copper, + eta_copper]) + + csvwriter.writerow(["", + "id_nitrogen", + "A_nitrogen (MHz)", + "euler_nitrogen (deg.)", + "NQCC_nitrogen (MHz)", + "eta_nitrogen", + "Cu_N_distance (Angstroms)"]) + + nitrogen_list = orcafile.find_element("N") + + for id_nitrogen in nitrogen_list: + atensor_nitrogen = cia.hyperfine(orcafile, id_nitrogen) + euler_nitrogen = cia.euler(orcafile, id_nitrogen) + nqi_nitrogen, nqcc_nitrogen, eta_nitrogen = cia.nqi(orcafile, id_nitrogen) + cu_n_dist = orcafile.pair_distance(id_copper, id_nitrogen) + + csvwriter.writerow(["", + id_nitrogen, + atensor_nitrogen, + euler_nitrogen, + nqcc_nitrogen, + eta_nitrogen, + cu_n_dist])
eb0772fc6c30d98b83bf1c8e7d83af21066ae45b
data_structures/Stack/Python/Stack.py
data_structures/Stack/Python/Stack.py
# Author: AlexBanks97 # Purpose: LIFO Stack implementation using python array. # Date: October 15th 2017 class Stack(object): def __init__(self): # Initialize stack as empty array self.stack = [] # Return and remove the last element of the stack array. def pop(self): # If the stack is not empty, pop. if self.stack.length > 0: return self.stack.pop() # Add an element to the end of the stack array. def push(self, element): self.stack.append(element)
# Author: AlexBanks97 # Purpose: LIFO Stack implementation using python array. # Date: October 15th 2017 class Stack(object): def __init__(self): # Initialize stack as empty array self.stack = [] # Return and remove the last element of the stack array. def pop(self): # If the stack is not empty, pop. if self.stack.length > 0: return self.stack.pop() # Add an element to the end of the stack array. def push(self, element): self.stack.append(element) # Return the last element of the stack array (without removing it). def peek(self): return self.stack[-1]
Add peek method and implementation
Add peek method and implementation
Python
cc0-1.0
Deepak345/al-go-rithms,Cnidarias/al-go-rithms,EUNIX-TRIX/al-go-rithms,ZoranPandovski/al-go-rithms,manikTharaka/al-go-rithms,Deepak345/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,Cnidarias/al-go-rithms,EUNIX-TRIX/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,EUNIX-TRIX/al-go-rithms,EUNIX-TRIX/al-go-rithms,manikTharaka/al-go-rithms,Cnidarias/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,Deepak345/al-go-rithms,Deepak345/al-go-rithms,Deepak345/al-go-rithms,Cnidarias/al-go-rithms,Cnidarias/al-go-rithms,Cnidarias/al-go-rithms,Cnidarias/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,manikTharaka/al-go-rithms,Deepak345/al-go-rithms,manikTharaka/al-go-rithms,Deepak345/al-go-rithms,ZoranPandovski/al-go-rithms,Deepak345/al-go-rithms,ZoranPandovski/al-go-rithms,Deepak345/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,EUNIX-TRIX/al-go-rithms,Deepak345/al-go-rithms,Deepak345/al-go-rithms,ZoranPandovski/al-go-rithms,Cnidarias/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,EUNIX-TRIX/al-go-rithms,Deepak345/al-go-rithms,Deepak345/al-go-rithms,manikTharaka/al-go-rithms,EUNIX-TRIX/al-go-rithms,ZoranPandovski/al-go-rithms,manikTharaka/al-go-rithms,Deepak345/al-go-rithms,ZoranPandovski/al-go-rithms,Deepak345/al-go-rithms,EUNIX-TRIX/al-go-rithms,EUNIX-TRIX/al-go-rithms,manikTharaka/al-go-rithms,EUNIX-TRIX/al-go-rithms,Cnidarias/al-go-rithms,Cnidarias/al-go-rithms,Cnidarias/al-go-rithms,manikTharaka/al-go-rithms,ZoranPandovski/al-go-rithms,Deepak345/al-go-rithms,manikTharaka/al-go-rithms,Cnidarias/al-go-rithms,manikTharaka/al-go-rithms,manikTharaka/al-go-rithms,Cnidarias/al-go-rithms,ZoranPandovski/al-go-rithms,EUNIX-TRIX/al-go-rithms,EUNIX-TRIX/al-go-rithms,Cnidarias/al-go-rithms,EUNIX-TRIX/al-go-rithms,ZoranPandovski/al-go-rithms,ZoranPandovski/al-go-rithms,Cnidarias/al-go-rithms
--- +++ @@ -16,3 +16,7 @@ # Add an element to the end of the stack array. def push(self, element): self.stack.append(element) + + # Return the last element of the stack array (without removing it). + def peek(self): + return self.stack[-1]
825c4d613915d43aea2e6ee0bc5d5b49ed0a4500
emission/analysis/classification/segmentation/section_segmentation.py
emission/analysis/classification/segmentation/section_segmentation.py
# Standard imports import attrdict as ad import numpy as np import datetime as pydt # Our imports import emission.analysis.classification.cleaning.location_smoothing as ls import emission.analysis.point_features as pf import emission.storage.decorations.location_queries as lq def segment_into_sections(trip): points_df = lq.get_activities_for_section(trip) no_tilting_points_df = points_df[points_df.activity != lq.Activities.TILTING] section_list = [] curr_section = ad.AttrDict({"user_id": trip.user_id, "loc_filter": trip.loc_filter, "start_ts": trip.start_ts, "start_time": trip.start_time, "activity": no_tilting_points_df.iloc[0].activity}) for idx, row in enumerate(no_tilting_points_df.to_dict('records')): if row["activity"] != curr_section.activity: # Let's add a second check here for confidence and types of activities if (row['agc'] > 60 and row['activity'] != lq.Activities.UNKNOWN and row['activity'] != lq.Activities.STILL): # Because the first section is initialized with the first activity. # So when idx == 0, the activities will be equal and this is # guaranteed to not be invoked assert(idx > 0) prev_ts = no_tilting_points_df.iloc[idx-1]["write_ts"] print("At %s, found new activity %s compared to current %s - creating new section with start_time %s" % (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), row["activity"], curr_section.activity, str(pydt.datetime.fromtimestamp(prev_ts/1000)))) # complete this section curr_section.end_ts = prev_ts curr_section.end_time = str(pydt.datetime.fromtimestamp(curr_section.end_ts/1000)) section_list.append(curr_section) # make a new section curr_section = ad.AttrDict({"user_id": trip.user_id, "loc_filter": trip.loc_filter, "start_ts": prev_ts, "start_time": pydt.datetime.fromtimestamp(prev_ts/1000), "activity": row["activity"]}) else: print("At %s, retained existing activity %s" % (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), curr_section.activity)) else: print("At %s, retained existing activity %s" % (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), curr_section.activity)) print("Detected trip end! Ending section at %s" % trip.end_time) # End the last section at the same time as the trip curr_section.end_ts = trip.end_ts curr_section.end_time = trip.end_time section_list.append(curr_section) return section_list
Create a simple method to segment a trip into sections
Create a simple method to segment a trip into sections This is purely based on the activity detection by android. It assumes that we have all activity updates. It uses a fairly naive algorithm with a threshold of 60% confidence, ignoring STILL, TILTING and UNKNOWN movements, and segmenting every time the mode changes. It seems to work pretty well for now, although I will tune it on an ongoing basis for the next few days.
Python
bsd-3-clause
shankari/e-mission-server,joshzarrabi/e-mission-server,joshzarrabi/e-mission-server,sunil07t/e-mission-server,yw374cornell/e-mission-server,e-mission/e-mission-server,e-mission/e-mission-server,yw374cornell/e-mission-server,yw374cornell/e-mission-server,sunil07t/e-mission-server,sunil07t/e-mission-server,shankari/e-mission-server,sunil07t/e-mission-server,yw374cornell/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,shankari/e-mission-server,joshzarrabi/e-mission-server,joshzarrabi/e-mission-server,e-mission/e-mission-server
--- +++ @@ -0,0 +1,55 @@ +# Standard imports +import attrdict as ad +import numpy as np +import datetime as pydt + +# Our imports +import emission.analysis.classification.cleaning.location_smoothing as ls +import emission.analysis.point_features as pf +import emission.storage.decorations.location_queries as lq + +def segment_into_sections(trip): + points_df = lq.get_activities_for_section(trip) + no_tilting_points_df = points_df[points_df.activity != lq.Activities.TILTING] + + section_list = [] + curr_section = ad.AttrDict({"user_id": trip.user_id, "loc_filter": trip.loc_filter, + "start_ts": trip.start_ts, "start_time": trip.start_time, + "activity": no_tilting_points_df.iloc[0].activity}) + for idx, row in enumerate(no_tilting_points_df.to_dict('records')): + if row["activity"] != curr_section.activity: + # Let's add a second check here for confidence and types of activities + if (row['agc'] > 60 and + row['activity'] != lq.Activities.UNKNOWN and + row['activity'] != lq.Activities.STILL): + # Because the first section is initialized with the first activity. + # So when idx == 0, the activities will be equal and this is + # guaranteed to not be invoked + assert(idx > 0) + prev_ts = no_tilting_points_df.iloc[idx-1]["write_ts"] + print("At %s, found new activity %s compared to current %s - creating new section with start_time %s" % + (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), + row["activity"], curr_section.activity, + str(pydt.datetime.fromtimestamp(prev_ts/1000)))) + # complete this section + curr_section.end_ts = prev_ts + curr_section.end_time = str(pydt.datetime.fromtimestamp(curr_section.end_ts/1000)) + section_list.append(curr_section) + # make a new section + curr_section = ad.AttrDict({"user_id": trip.user_id, "loc_filter": trip.loc_filter, + "start_ts": prev_ts, + "start_time": pydt.datetime.fromtimestamp(prev_ts/1000), + "activity": row["activity"]}) + else: + print("At %s, retained existing activity %s" % + (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), curr_section.activity)) + else: + print("At %s, retained existing activity %s" % + (str(pydt.datetime.fromtimestamp(row["write_ts"]/1000)), curr_section.activity)) + + print("Detected trip end! Ending section at %s" % trip.end_time) + # End the last section at the same time as the trip + curr_section.end_ts = trip.end_ts + curr_section.end_time = trip.end_time + section_list.append(curr_section) + return section_list
5f2cd26054adff5a1fbf9ba5d56766b972f46670
leakcheck/thread-key-gen.py
leakcheck/thread-key-gen.py
# Copyright (C) Jean-Paul Calderone # See LICENSE for details. # # Stress tester for thread-related bugs in RSA and DSA key generation. 0.12 and # older held the GIL during these operations. Subsequent versions release it # during them. from threading import Thread from OpenSSL.crypto import TYPE_RSA, TYPE_DSA, PKey def generate_rsa(): keys = [] for i in range(100): key = PKey() key.generate_key(TYPE_RSA, 1024) keys.append(key) def generate_dsa(): keys = [] for i in range(100): key = PKey() key.generate_key(TYPE_DSA, 512) keys.append(key) def main(): threads = [] for i in range(3): t = Thread(target=generate_rsa, args=()) threads.append(t) t = Thread(target=generate_dsa, args=()) threads.append(t) for t in threads: t.start() main()
Add a multithreaded stress tester for key generation. Hopefully provides additional confidence that that code is correct with respect to threading.
Add a multithreaded stress tester for key generation. Hopefully provides additional confidence that that code is correct with respect to threading.
Python
apache-2.0
mhils/pyopenssl,kediacorporation/pyopenssl,reaperhulk/pyopenssl,r0ro/pyopenssl,mhils/pyopenssl,daodaoliang/pyopenssl,elitest/pyopenssl,kjav/pyopenssl,kediacorporation/pyopenssl,alex/pyopenssl,lvh/pyopenssl,samv/pyopenssl,Lukasa/pyopenssl,mitghi/pyopenssl,msabramo/pyOpenSSL,reaperhulk/pyopenssl,hynek/pyopenssl,pyca/pyopenssl,rackerlabs/pyopenssl,rackerlabs/pyopenssl,EnerNOC/pyopenssl,lvh/pyopenssl,adamwolf/pyopenssl,Lukasa/pyopenssl,mitghi/pyopenssl,sorenh/pyopenssl,EnerNOC/pyopenssl,msabramo/pyOpenSSL,EnerNOC/pyopenssl,hynek/pyopenssl,msabramo/pyOpenSSL,aalba6675/pyopenssl,mschmo/pyopenssl,aalba6675/pyopenssl,JensTimmerman/pyopenssl,r0ro/pyopenssl,sholsapp/pyopenssl,alex/pyopenssl,JensTimmerman/pyopenssl
--- +++ @@ -0,0 +1,38 @@ +# Copyright (C) Jean-Paul Calderone +# See LICENSE for details. +# +# Stress tester for thread-related bugs in RSA and DSA key generation. 0.12 and +# older held the GIL during these operations. Subsequent versions release it +# during them. + +from threading import Thread + +from OpenSSL.crypto import TYPE_RSA, TYPE_DSA, PKey + +def generate_rsa(): + keys = [] + for i in range(100): + key = PKey() + key.generate_key(TYPE_RSA, 1024) + keys.append(key) + +def generate_dsa(): + keys = [] + for i in range(100): + key = PKey() + key.generate_key(TYPE_DSA, 512) + keys.append(key) + + +def main(): + threads = [] + for i in range(3): + t = Thread(target=generate_rsa, args=()) + threads.append(t) + t = Thread(target=generate_dsa, args=()) + threads.append(t) + + for t in threads: + t.start() + +main()
4820013e207947fe7ff94777cd8dcf1ed474eab1
migrations/versions/fb6a6554b21_add_account_lockout_fields_to_user.py
migrations/versions/fb6a6554b21_add_account_lockout_fields_to_user.py
"""Add account lockout fields to User Revision ID: fb6a6554b21 Revises: 1f9b411bf6df Create Date: 2015-10-29 01:07:27.930095 """ # revision identifiers, used by Alembic. revision = 'fb6a6554b21' down_revision = '1f9b411bf6df' from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### op.add_column('users', sa.Column('failed_login_attempts', sa.Integer(), nullable=True)) op.add_column('users', sa.Column('last_failed_login_attempt', sa.DateTime(), nullable=True)) op.add_column('users', sa.Column('locked_out', sa.Boolean(), nullable=True)) ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### op.drop_column('users', 'locked_out') op.drop_column('users', 'last_failed_login_attempt') op.drop_column('users', 'failed_login_attempts') ### end Alembic commands ###
Add migration for account lockout fields on User
Add migration for account lockout fields on User
Python
mit
richgieg/flask-now,richgieg/flask-now
--- +++ @@ -0,0 +1,30 @@ +"""Add account lockout fields to User + +Revision ID: fb6a6554b21 +Revises: 1f9b411bf6df +Create Date: 2015-10-29 01:07:27.930095 + +""" + +# revision identifiers, used by Alembic. +revision = 'fb6a6554b21' +down_revision = '1f9b411bf6df' + +from alembic import op +import sqlalchemy as sa + + +def upgrade(): + ### commands auto generated by Alembic - please adjust! ### + op.add_column('users', sa.Column('failed_login_attempts', sa.Integer(), nullable=True)) + op.add_column('users', sa.Column('last_failed_login_attempt', sa.DateTime(), nullable=True)) + op.add_column('users', sa.Column('locked_out', sa.Boolean(), nullable=True)) + ### end Alembic commands ### + + +def downgrade(): + ### commands auto generated by Alembic - please adjust! ### + op.drop_column('users', 'locked_out') + op.drop_column('users', 'last_failed_login_attempt') + op.drop_column('users', 'failed_login_attempts') + ### end Alembic commands ###
6fabbe85bb74788641897daf8b162eac3d47b0aa
data_crunching/indonesia_timeseries/download_indonesia_prices.py
data_crunching/indonesia_timeseries/download_indonesia_prices.py
#!/usr/bin/env python2 import urllib2 import shutil import re import sys import datetime from lxml import etree usage_str = """ This scripts downloads daily food prices from http://m.pip.kementan.org/index.php (Indonesia). Provide date in DD/MM/YYYY format. Example: ./download_indonesia_prices.py 15/03/2013 """ def download_table(date): """Download price table for a given date""" main_url = 'http://m.pip.kementan.org/index.php' params = 'laporan=LHK-01&tanggal=%s&bulan=%s&tahun=%s&pilihlaporan=View+Laporan' % (date.day, date.month, date.year) req = urllib2.Request(main_url, params) response = urllib2.urlopen(req) html_code = response.read() regex = re.compile(r'<div id="content" align="center">.*(<table.+</table>)', re.DOTALL) match = regex.search(html_code) if not match: print "ERROR: table not detected" sys.exit(1) table_html = match.group(1) # Remove commas table_html = re.sub(r'(?<=\d),(?=\d)', '', table_html) table = etree.XML(table_html) rows = iter(table) actual_headers = [col.text for col in next(rows)] # TODO: translate this bullshit ;) headers = ['Dried Grain Harvest', 'Dry unhusked', 'Rice Medium', 'Rice Premium', 'Corn', 'Local soybean', 'Local Peanuts', 'Green Beans', 'Cassava', 'Sweet potato', 'Cassava spindles'] print "; ".join(headers), "\n" # Print table for row in rows: if all(v.text is None for v in row): continue print ('''"%s"''') % row[0].text, for col in row[1:]: print col.text, print def parse_date(date_string): """Check date""" match = re.match(r'(\d{2})/(\d{2})/(\d{4})', date_string) if not match: sys.exit("ERROR: invalid date") day, month, year = int(match.group(1)), int(match.group(2)), int(match.group(3)) return datetime.date(year, month, day) def usage(): print usage_str if __name__ == "__main__": if len(sys.argv) == 1 or sys.argv[1] in ['-h', '--help']: usage() sys.exit(0) date_string = sys.argv[1] date = parse_date(date_string) download_table(date)
Add script for downloading Indonesia price data
Add script for downloading Indonesia price data (http://m.pip.kementan.org/index.php)
Python
bsd-3-clause
FAB4D/humanitas,FAB4D/humanitas,FAB4D/humanitas
--- +++ @@ -0,0 +1,78 @@ +#!/usr/bin/env python2 + +import urllib2 +import shutil +import re +import sys +import datetime +from lxml import etree + +usage_str = """ +This scripts downloads daily food prices from http://m.pip.kementan.org/index.php (Indonesia). + +Provide date in DD/MM/YYYY format. + +Example: + + ./download_indonesia_prices.py 15/03/2013 +""" + +def download_table(date): + """Download price table for a given date""" + + main_url = 'http://m.pip.kementan.org/index.php' + + params = 'laporan=LHK-01&tanggal=%s&bulan=%s&tahun=%s&pilihlaporan=View+Laporan' % (date.day, date.month, date.year) + + req = urllib2.Request(main_url, params) + response = urllib2.urlopen(req) + html_code = response.read() + + regex = re.compile(r'<div id="content" align="center">.*(<table.+</table>)', re.DOTALL) + match = regex.search(html_code) + if not match: + print "ERROR: table not detected" + sys.exit(1) + table_html = match.group(1) + + # Remove commas + table_html = re.sub(r'(?<=\d),(?=\d)', '', table_html) + + table = etree.XML(table_html) + rows = iter(table) + actual_headers = [col.text for col in next(rows)] + # TODO: translate this bullshit ;) + headers = ['Dried Grain Harvest', 'Dry unhusked', 'Rice Medium', 'Rice Premium', 'Corn', 'Local soybean', 'Local Peanuts', 'Green Beans', 'Cassava', 'Sweet potato', 'Cassava spindles'] + print "; ".join(headers), "\n" + + + # Print table + for row in rows: + if all(v.text is None for v in row): + continue + print ('''"%s"''') % row[0].text, + for col in row[1:]: + print col.text, + print + +def parse_date(date_string): + """Check date""" + + match = re.match(r'(\d{2})/(\d{2})/(\d{4})', date_string) + if not match: + sys.exit("ERROR: invalid date") + day, month, year = int(match.group(1)), int(match.group(2)), int(match.group(3)) + return datetime.date(year, month, day) + +def usage(): + print usage_str + +if __name__ == "__main__": + if len(sys.argv) == 1 or sys.argv[1] in ['-h', '--help']: + usage() + sys.exit(0) + + date_string = sys.argv[1] + date = parse_date(date_string) + download_table(date) +
a795d94a9c885b97ab5bffc313524ae46626d556
tools/analyze_code_size.py
tools/analyze_code_size.py
import os import re import sys import optparse MARKER_START_FUNCS = "// EMSCRIPTEN_START_FUNCS" MARKER_END_FUNCS = "// EMSCRIPTEN_END_FUNCS" FUNCTION_CODE_RE = re.compile( r"function (?P<name>[a-zA-Z0-9_]+)(?P<defn>.*?)((?=function)|(?=$))" ) def analyze_code_size(fileobj, opts): funcs = {} name_re = None if opts.grep is not None: name_re = re.compile(opts.grep, re.I) # Split out and analyze the code for each individual function. # XXX TODO: read incrementally to reduce memory usage. data = fileobj.read() pre_code, data = data.split(MARKER_START_FUNCS, 1) data, post_code = data.split(MARKER_END_FUNCS, 1) for match in FUNCTION_CODE_RE.finditer(data): name = match.group("name") defn = match.group("defn") if name_re and not name_re.search(name): continue funcs[name] = FunctionMetrics(name, defn) # Print summary metrics. total = 0 funcs_by_size = ((f.size, f.name) for f in funcs.itervalues()) for (size, name) in sorted(funcs_by_size, reverse=True): print size, name, human_readable(size) total += size print "Total size:", total, human_readable(total) class FunctionMetrics(object): def __init__(self, name, defn): self.name = name self.defn = defn self.size = len(defn) def human_readable(size): units = ((1024*1024, "M"), (1024, "k")) for (scale, unit) in units: scale = float(scale) if size / scale > 0.1: return "(%.2f%s)" % (size / scale, unit) return "" def main(args=None): usage = "usage: %prog [options] file" descr = "Analyze code size and complexity for emscripten-compiled output" parser = optparse.OptionParser(usage=usage, description=descr) parser.add_option("-g", "--grep", metavar="REGEXP", help="only analyze functions matching this regexp") opts, args = parser.parse_args(args) with open(args[0], "r") as infile: analyze_code_size(infile, opts) return 0 if __name__ == "__main__": try: exitcode = main() except KeyboardInterrupt: exitcode = 1 sys.exit(exitcode)
Add simple function-size analysis tool.
Add simple function-size analysis tool.
Python
mit
pypyjs/pypyjs,perkinslr/pypyjs,perkinslr/pypyjs,perkinslr/pypyjs,albertjan/pypyjs,pombredanne/pypyjs,albertjan/pypyjs,perkinslr/pypyjs,pombredanne/pypyjs,pypyjs/pypyjs,trinketapp/pypyjs,pypyjs/pypyjs,pypyjs/pypyjs,perkinslr/pypyjs,trinketapp/pypyjs,albertjan/pypyjs
--- +++ @@ -0,0 +1,76 @@ + +import os +import re +import sys +import optparse + +MARKER_START_FUNCS = "// EMSCRIPTEN_START_FUNCS" +MARKER_END_FUNCS = "// EMSCRIPTEN_END_FUNCS" + +FUNCTION_CODE_RE = re.compile( + r"function (?P<name>[a-zA-Z0-9_]+)(?P<defn>.*?)((?=function)|(?=$))" +) + + +def analyze_code_size(fileobj, opts): + funcs = {} + name_re = None + if opts.grep is not None: + name_re = re.compile(opts.grep, re.I) + # Split out and analyze the code for each individual function. + # XXX TODO: read incrementally to reduce memory usage. + data = fileobj.read() + pre_code, data = data.split(MARKER_START_FUNCS, 1) + data, post_code = data.split(MARKER_END_FUNCS, 1) + for match in FUNCTION_CODE_RE.finditer(data): + name = match.group("name") + defn = match.group("defn") + if name_re and not name_re.search(name): + continue + funcs[name] = FunctionMetrics(name, defn) + # Print summary metrics. + total = 0 + funcs_by_size = ((f.size, f.name) for f in funcs.itervalues()) + for (size, name) in sorted(funcs_by_size, reverse=True): + print size, name, human_readable(size) + total += size + print "Total size:", total, human_readable(total) + + +class FunctionMetrics(object): + + def __init__(self, name, defn): + self.name = name + self.defn = defn + self.size = len(defn) + + +def human_readable(size): + units = ((1024*1024, "M"), (1024, "k")) + for (scale, unit) in units: + scale = float(scale) + if size / scale > 0.1: + return "(%.2f%s)" % (size / scale, unit) + return "" + + + +def main(args=None): + usage = "usage: %prog [options] file" + descr = "Analyze code size and complexity for emscripten-compiled output" + parser = optparse.OptionParser(usage=usage, description=descr) + parser.add_option("-g", "--grep", metavar="REGEXP", + help="only analyze functions matching this regexp") + + opts, args = parser.parse_args(args) + with open(args[0], "r") as infile: + analyze_code_size(infile, opts) + return 0 + + +if __name__ == "__main__": + try: + exitcode = main() + except KeyboardInterrupt: + exitcode = 1 + sys.exit(exitcode)
bfaeeec3f5f5582822e2918491090815a606ba44
test/test_api.py
test/test_api.py
# -*- coding: utf-8 -*- import warthog.api def test_public_exports(): exports = set([item for item in dir(warthog.api) if not item.startswith('_')]) declared = set(warthog.api.__all__) assert exports == declared, 'Exports and __all__ members should match'
Add test to make sure imports and __all__ matches
Add test to make sure imports and __all__ matches
Python
mit
smarter-travel-media/warthog
--- +++ @@ -0,0 +1,9 @@ +# -*- coding: utf-8 -*- + +import warthog.api + + +def test_public_exports(): + exports = set([item for item in dir(warthog.api) if not item.startswith('_')]) + declared = set(warthog.api.__all__) + assert exports == declared, 'Exports and __all__ members should match'
48857638694ceca08c64d7b9c6825e2178c53279
pylearn2/utils/doc.py
pylearn2/utils/doc.py
""" Documentation-related helper classes/functions """ class soft_wraps: """ A Python decorator which concatenates two functions' docstrings: one function is defined at initialization and the other one is defined when soft_wraps is called. This helps reduce the ammount of documentation to write: one can use this decorator on child classes' functions when their implementation is similar to the one of the parent class. Conversely, if a function defined in a child class departs from its parent's implementation, one can simply explain the differences in a 'Notes' section without re-writing the whole docstring. Examples -------- >>> class Parent(object): ... def f(x): ... ''' ... Adds 1 to x ... ... Parameters ... ---------- ... x : int ... Variable to increment by 1 ... ... Returns ... ------- ... rval : int ... x incremented by 1 ... ''' ... rval = x + 1 ... return rval ... >>> class Child(Parent): ... @soft_wraps(Parent.f) ... def f(x): ... ''' ... Notes ... ----- ... Also prints the incremented value ... ''' ... rval = x + 1 ... print rval ... return rval ... >>> c = Child() >>> print c.f.__doc__ Adds 1 to x Parameters ---------- x : int Variable to increment by 1 Returns ------- rval : int x incremented by 1 Notes ----- Also prints the incremented value """ def __init__(self, f, append=False): """ Parameters ---------- f : function Function whose docstring will be concatenated with the decorated function's docstring prepend : bool, optional If True, appends f's docstring to the decorated function's docstring instead of prepending it. Defaults to False. """ self.f = f self.append = append def __call__(self, f): """ Prepend self.f's docstring to f's docstring (or append it if `self.append == True`). Parameters ---------- f : function Function to decorate Returns ------- f : function Function f passed as argument with self.f's docstring {pre,ap}pended to it """ if self.append: f.__doc__ += + self.f.__doc__ else: f.__doc__ = self.f.__doc__ + f.__doc__ return f
Add function decorator to improve functools.wraps
Add function decorator to improve functools.wraps
Python
bsd-3-clause
goodfeli/pylearn2,JesseLivezey/pylearn2,TNick/pylearn2,fulmicoton/pylearn2,pkainz/pylearn2,Refefer/pylearn2,woozzu/pylearn2,kastnerkyle/pylearn2,CIFASIS/pylearn2,mclaughlin6464/pylearn2,hyqneuron/pylearn2-maxsom,aalmah/pylearn2,bartvm/pylearn2,JesseLivezey/pylearn2,nouiz/pylearn2,lamblin/pylearn2,CIFASIS/pylearn2,junbochen/pylearn2,ddboline/pylearn2,junbochen/pylearn2,ddboline/pylearn2,alexjc/pylearn2,w1kke/pylearn2,abergeron/pylearn2,mkraemer67/pylearn2,jamessergeant/pylearn2,fyffyt/pylearn2,fishcorn/pylearn2,fyffyt/pylearn2,matrogers/pylearn2,matrogers/pylearn2,aalmah/pylearn2,lunyang/pylearn2,skearnes/pylearn2,mkraemer67/pylearn2,theoryno3/pylearn2,kose-y/pylearn2,se4u/pylearn2,aalmah/pylearn2,daemonmaker/pylearn2,jeremyfix/pylearn2,hyqneuron/pylearn2-maxsom,lancezlin/pylearn2,Refefer/pylearn2,ddboline/pylearn2,lancezlin/pylearn2,kose-y/pylearn2,JesseLivezey/plankton,hantek/pylearn2,goodfeli/pylearn2,woozzu/pylearn2,ashhher3/pylearn2,bartvm/pylearn2,shiquanwang/pylearn2,TNick/pylearn2,hantek/pylearn2,lancezlin/pylearn2,TNick/pylearn2,daemonmaker/pylearn2,pkainz/pylearn2,fyffyt/pylearn2,alexjc/pylearn2,lunyang/pylearn2,ddboline/pylearn2,hantek/pylearn2,woozzu/pylearn2,pombredanne/pylearn2,TNick/pylearn2,jeremyfix/pylearn2,hyqneuron/pylearn2-maxsom,fishcorn/pylearn2,lisa-lab/pylearn2,kastnerkyle/pylearn2,mkraemer67/pylearn2,abergeron/pylearn2,alexjc/pylearn2,pombredanne/pylearn2,hyqneuron/pylearn2-maxsom,daemonmaker/pylearn2,KennethPierce/pylearnk,cosmoharrigan/pylearn2,aalmah/pylearn2,shiquanwang/pylearn2,JesseLivezey/plankton,KennethPierce/pylearnk,jamessergeant/pylearn2,caidongyun/pylearn2,fishcorn/pylearn2,mkraemer67/pylearn2,jamessergeant/pylearn2,CIFASIS/pylearn2,chrish42/pylearn,lunyang/pylearn2,fulmicoton/pylearn2,bartvm/pylearn2,mclaughlin6464/pylearn2,lunyang/pylearn2,se4u/pylearn2,fyffyt/pylearn2,Refefer/pylearn2,matrogers/pylearn2,mclaughlin6464/pylearn2,se4u/pylearn2,nouiz/pylearn2,jamessergeant/pylearn2,JesseLivezey/pylearn2,chrish42/pylearn,KennethPierce/pylearnk,cosmoharrigan/pylearn2,sandeepkbhat/pylearn2,theoryno3/pylearn2,w1kke/pylearn2,goodfeli/pylearn2,JesseLivezey/pylearn2,Refefer/pylearn2,caidongyun/pylearn2,msingh172/pylearn2,kastnerkyle/pylearn2,skearnes/pylearn2,KennethPierce/pylearnk,abergeron/pylearn2,kastnerkyle/pylearn2,skearnes/pylearn2,jeremyfix/pylearn2,pombredanne/pylearn2,ashhher3/pylearn2,lisa-lab/pylearn2,lamblin/pylearn2,junbochen/pylearn2,ashhher3/pylearn2,lamblin/pylearn2,CIFASIS/pylearn2,lisa-lab/pylearn2,pkainz/pylearn2,pombredanne/pylearn2,fishcorn/pylearn2,se4u/pylearn2,pkainz/pylearn2,junbochen/pylearn2,cosmoharrigan/pylearn2,goodfeli/pylearn2,theoryno3/pylearn2,ashhher3/pylearn2,alexjc/pylearn2,jeremyfix/pylearn2,JesseLivezey/plankton,chrish42/pylearn,msingh172/pylearn2,mclaughlin6464/pylearn2,sandeepkbhat/pylearn2,shiquanwang/pylearn2,w1kke/pylearn2,caidongyun/pylearn2,skearnes/pylearn2,matrogers/pylearn2,fulmicoton/pylearn2,shiquanwang/pylearn2,chrish42/pylearn,lisa-lab/pylearn2,kose-y/pylearn2,hantek/pylearn2,kose-y/pylearn2,msingh172/pylearn2,sandeepkbhat/pylearn2,w1kke/pylearn2,msingh172/pylearn2,lancezlin/pylearn2,woozzu/pylearn2,daemonmaker/pylearn2,JesseLivezey/plankton,fulmicoton/pylearn2,caidongyun/pylearn2,bartvm/pylearn2,nouiz/pylearn2,cosmoharrigan/pylearn2,sandeepkbhat/pylearn2,nouiz/pylearn2,lamblin/pylearn2,theoryno3/pylearn2,abergeron/pylearn2
--- +++ @@ -0,0 +1,106 @@ +""" +Documentation-related helper classes/functions +""" + + +class soft_wraps: + """ + A Python decorator which concatenates two functions' docstrings: one + function is defined at initialization and the other one is defined when + soft_wraps is called. + + This helps reduce the ammount of documentation to write: one can use + this decorator on child classes' functions when their implementation is + similar to the one of the parent class. Conversely, if a function defined + in a child class departs from its parent's implementation, one can simply + explain the differences in a 'Notes' section without re-writing the whole + docstring. + + Examples + -------- + >>> class Parent(object): + ... def f(x): + ... ''' + ... Adds 1 to x + ... + ... Parameters + ... ---------- + ... x : int + ... Variable to increment by 1 + ... + ... Returns + ... ------- + ... rval : int + ... x incremented by 1 + ... ''' + ... rval = x + 1 + ... return rval + ... + >>> class Child(Parent): + ... @soft_wraps(Parent.f) + ... def f(x): + ... ''' + ... Notes + ... ----- + ... Also prints the incremented value + ... ''' + ... rval = x + 1 + ... print rval + ... return rval + ... + >>> c = Child() + >>> print c.f.__doc__ + + Adds 1 to x + + Parameters + ---------- + x : int + Variable to increment by 1 + + Returns + ------- + rval : int + x incremented by 1 + + Notes + ----- + Also prints the incremented value + """ + + def __init__(self, f, append=False): + """ + Parameters + ---------- + f : function + Function whose docstring will be concatenated with the decorated + function's docstring + prepend : bool, optional + If True, appends f's docstring to the decorated function's + docstring instead of prepending it. Defaults to False. + """ + self.f = f + self.append = append + + def __call__(self, f): + """ + Prepend self.f's docstring to f's docstring (or append it if + `self.append == True`). + + Parameters + ---------- + f : function + Function to decorate + + Returns + ------- + f : function + Function f passed as argument with self.f's docstring + {pre,ap}pended to it + """ + if self.append: + f.__doc__ += + self.f.__doc__ + else: + f.__doc__ = self.f.__doc__ + f.__doc__ + + return f
7ff614950163b1fb6a8fe0fef5b8de9bfa3a9d85
transmutagen/tests/test_partialfrac.py
transmutagen/tests/test_partialfrac.py
from sympy import together, expand_complex, re, im, symbols from ..partialfrac import t def test_re_form(): theta, alpha = symbols('theta, alpha') # Check that this doesn't change re_form = together(expand_complex(re(alpha/(t - theta)))) assert re_form == (t*re(alpha) - re(alpha)*re(theta) - im(alpha)*im(theta))/((t - re(theta))**2 + im(theta)**2)
Add a test for the hard-coded re() partial frac form
Add a test for the hard-coded re() partial frac form
Python
bsd-3-clause
ergs/transmutagen,ergs/transmutagen
--- +++ @@ -0,0 +1,11 @@ +from sympy import together, expand_complex, re, im, symbols + +from ..partialfrac import t + +def test_re_form(): + theta, alpha = symbols('theta, alpha') + + # Check that this doesn't change + re_form = together(expand_complex(re(alpha/(t - theta)))) + assert re_form == (t*re(alpha) - re(alpha)*re(theta) - + im(alpha)*im(theta))/((t - re(theta))**2 + im(theta)**2)
8dc7a1e239dc22dd4eb69cfe1754586e3a1690dc
tests/test_run_js.py
tests/test_run_js.py
import os from py2js import JavaScript def f(x): return x def test(func, run): func_source = str(JavaScript(func)) run_file = "/tmp/run.js" with open(run_file, "w") as f: f.write(func_source) f.write("\n") f.write(run) r = os.system('js -f defs.js -f %s' % run_file) assert r == 0 test(f, "assert(f(3) == 3)") test(f, "assert(f(3) != 4)")
Test javascript using the "js"
Test javascript using the "js"
Python
mit
qsnake/py2js,mattpap/py2js,chrivers/pyjaco,buchuki/pyjaco,mattpap/py2js,chrivers/pyjaco,chrivers/pyjaco,buchuki/pyjaco,buchuki/pyjaco,qsnake/py2js
--- +++ @@ -0,0 +1,20 @@ +import os + +from py2js import JavaScript + +def f(x): + return x + +def test(func, run): + func_source = str(JavaScript(func)) + run_file = "/tmp/run.js" + with open(run_file, "w") as f: + f.write(func_source) + f.write("\n") + f.write(run) + r = os.system('js -f defs.js -f %s' % run_file) + assert r == 0 + + +test(f, "assert(f(3) == 3)") +test(f, "assert(f(3) != 4)")
3dd71c02ea1fa9e39054bd82bf9e8657ec77d6b9
tools/get_chat_id.py
tools/get_chat_id.py
#! /usr/bin/python3 # -*- coding:utf-8 -*- # by antoine@2ohm.fr import sys import time import telepot def handle(msg): content_type, chat_type, chat_id = telepot.glance(msg) print("\tchat_id: {}".format(chat_id)) if content_type == 'text' and msg['text'] == '/start': ans = """ Hello <b>{first_name}</b>, nice to meet you!\n Your chat_id is <code>{chat_id}</code>.\n You can stop the <code>get_chat_id</code> script with <code>CTRL+C</code> and start using the ProgressBot right now.\n See you soon! """.format(first_name = msg['from']['first_name'], chat_id = chat_id) bot.sendMessage(chat_id, ans, parse_mode = "HTML") TOKEN = "PUT_YOUR_TOKKEN_HERE" bot = telepot.Bot(TOKEN) bot.message_loop(handle) print ('Listening ...') # Keep the program running. while 1: try: time.sleep(10) except KeyboardInterrupt: print() sys.exit()
Add a script to recover the chat_id
Add a script to recover the chat_id
Python
apache-2.0
a2ohm/ProgressBot
--- +++ @@ -0,0 +1,37 @@ +#! /usr/bin/python3 +# -*- coding:utf-8 -*- + +# by antoine@2ohm.fr + +import sys +import time +import telepot + +def handle(msg): + content_type, chat_type, chat_id = telepot.glance(msg) + print("\tchat_id: {}".format(chat_id)) + + if content_type == 'text' and msg['text'] == '/start': + ans = """ +Hello <b>{first_name}</b>, nice to meet you!\n +Your chat_id is <code>{chat_id}</code>.\n +You can stop the <code>get_chat_id</code> script with <code>CTRL+C</code> and start using the ProgressBot right now.\n +See you soon! + """.format(first_name = msg['from']['first_name'], + chat_id = chat_id) + + bot.sendMessage(chat_id, ans, parse_mode = "HTML") + +TOKEN = "PUT_YOUR_TOKKEN_HERE" + +bot = telepot.Bot(TOKEN) +bot.message_loop(handle) +print ('Listening ...') + +# Keep the program running. +while 1: + try: + time.sleep(10) + except KeyboardInterrupt: + print() + sys.exit()
fed2e3f9bdb3a00b077b5e7df1aed4d927b77b6c
tests/clifford_test.py
tests/clifford_test.py
"""Test for the Clifford algebra drudge.""" from drudge import CliffordDrudge, Vec, inner_by_delta def test_clifford_drudge_by_quaternions(spark_ctx): """Test basic functionality of Clifford drudge by quaternions. """ dr = CliffordDrudge( spark_ctx, inner=lambda v1, v2: -inner_by_delta(v1, v2) ) e_ = Vec('e') i_ = dr.sum(e_[2] * e_[3]).simplify() j_ = dr.sum(e_[3] * e_[1]).simplify() k_ = dr.sum(e_[1] * e_[2]).simplify() for i in [i_, j_, k_]: assert (i * i).simplify() == -1 assert (i_ * j_ * k_).simplify() == -1 assert (i_ * j_).simplify() == k_ assert (j_ * k_).simplify() == i_ assert (k_ * i_).simplify() == j_
Add test for Clifford drudge by quaternions
Add test for Clifford drudge by quaternions
Python
mit
tschijnmo/drudge,tschijnmo/drudge,tschijnmo/drudge
--- +++ @@ -0,0 +1,26 @@ +"""Test for the Clifford algebra drudge.""" + +from drudge import CliffordDrudge, Vec, inner_by_delta + + +def test_clifford_drudge_by_quaternions(spark_ctx): + """Test basic functionality of Clifford drudge by quaternions. + """ + + dr = CliffordDrudge( + spark_ctx, inner=lambda v1, v2: -inner_by_delta(v1, v2) + ) + e_ = Vec('e') + + i_ = dr.sum(e_[2] * e_[3]).simplify() + j_ = dr.sum(e_[3] * e_[1]).simplify() + k_ = dr.sum(e_[1] * e_[2]).simplify() + + for i in [i_, j_, k_]: + assert (i * i).simplify() == -1 + + assert (i_ * j_ * k_).simplify() == -1 + + assert (i_ * j_).simplify() == k_ + assert (j_ * k_).simplify() == i_ + assert (k_ * i_).simplify() == j_
09a0689b8e521c1d5c0ea68ac448dc9ae7abcff5
fitsHeader.py
fitsHeader.py
#!/usr/bin/env python # -*- coding: utf8 -*- # My imports from __future__ import division from astropy.io import fits from pydoc import pager import argparse def _parser(): parser = argparse.ArgumentParser(description='View the header of a fits file') parser.add_argument('input', help='File name of fits file') parser.add_argument('-key', help='Look up a given key (case insensitive)', default=None) return parser.parse_args() if __name__ == '__main__': args = _parser() h = fits.getheader(args.input) h.keys = map(str.lower, h.keys()) if args.key: args.key = args.key.lower() try: print h[args.key] except KeyError: raise KeyError('Key was not found') else: string = '\n'.join("{!s} : {!r}".format(key, val) for (key, val) in h.items()) pager(string)
Read the header of a fits file and/or look up a single key (case insensitive).
Read the header of a fits file and/or look up a single key (case insensitive).
Python
mit
DanielAndreasen/astro_scripts
--- +++ @@ -0,0 +1,31 @@ +#!/usr/bin/env python +# -*- coding: utf8 -*- + +# My imports +from __future__ import division +from astropy.io import fits +from pydoc import pager +import argparse + + +def _parser(): + parser = argparse.ArgumentParser(description='View the header of a fits file') + parser.add_argument('input', help='File name of fits file') + parser.add_argument('-key', help='Look up a given key (case insensitive)', default=None) + return parser.parse_args() + + +if __name__ == '__main__': + args = _parser() + h = fits.getheader(args.input) + h.keys = map(str.lower, h.keys()) + + if args.key: + args.key = args.key.lower() + try: + print h[args.key] + except KeyError: + raise KeyError('Key was not found') + else: + string = '\n'.join("{!s} : {!r}".format(key, val) for (key, val) in h.items()) + pager(string)
b674f921a8e5cffb2d3e320f564c61ca01455a9f
wafer/management/commands/wafer_talk_video_reviewers.py
wafer/management/commands/wafer_talk_video_reviewers.py
import sys import csv from django.core.management.base import BaseCommand from django.contrib.auth import get_user_model from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL class Command(BaseCommand): help = ("List talks and the associated video_reviewer emails." " Only reviewers for accepted talks are listed") def _video_reviewers(self, options): talks = Talk.objects.filter(status=ACCEPTED) csv_file = csv.writer(sys.stdout) for talk in talks: reviewer = talk.video_reviewer if not reviewer: reviewer = 'NO REVIEWER' row = [x.encode("utf-8") for x in ( talk.title, reviewer, )] csv_file.writerow(row) def handle(self, *args, **options): self._video_reviewers(options)
Add command to generate a csv of talk titles and video reviewers
Add command to generate a csv of talk titles and video reviewers
Python
isc
CTPUG/wafer,CTPUG/wafer,CTPUG/wafer,CTPUG/wafer
--- +++ @@ -0,0 +1,29 @@ +import sys +import csv + +from django.core.management.base import BaseCommand + +from django.contrib.auth import get_user_model +from wafer.talks.models import Talk, ACCEPTED, PROVISIONAL + + +class Command(BaseCommand): + help = ("List talks and the associated video_reviewer emails." + " Only reviewers for accepted talks are listed") + + def _video_reviewers(self, options): + talks = Talk.objects.filter(status=ACCEPTED) + + csv_file = csv.writer(sys.stdout) + for talk in talks: + reviewer = talk.video_reviewer + if not reviewer: + reviewer = 'NO REVIEWER' + row = [x.encode("utf-8") for x in ( + talk.title, + reviewer, + )] + csv_file.writerow(row) + + def handle(self, *args, **options): + self._video_reviewers(options)
3db3c22d83071550d8bbd70062f957cf43c5e54a
cart/_compatibility.py
cart/_compatibility.py
import sys is_py3 = sys.version_info[0] >= 3 def utf8(string): """Cast to unicode DAMMIT! Written because Python2 repr always implicitly casts to a string, so we have to cast back to a unicode (and we now that we always deal with valid unicode, because we check that in the beginning). """ if is_py3: return str(string) elif not isinstance(string, unicode): return unicode(str(string), 'UTF-8') return string
Add a compatibility module, because of Python 2/3 compatibility issues.
Add a compatibility module, because of Python 2/3 compatibility issues.
Python
mit
davidhalter-archive/shopping_cart_example
--- +++ @@ -0,0 +1,16 @@ +import sys + +is_py3 = sys.version_info[0] >= 3 + + +def utf8(string): + """Cast to unicode DAMMIT! + Written because Python2 repr always implicitly casts to a string, so we + have to cast back to a unicode (and we now that we always deal with valid + unicode, because we check that in the beginning). + """ + if is_py3: + return str(string) + elif not isinstance(string, unicode): + return unicode(str(string), 'UTF-8') + return string
156b7dfc11f24a7d77d2280e8ddade3cb7a474b7
misc/list_all_es_indexes.py
misc/list_all_es_indexes.py
#!/usr/bin/env python # -*- encoding: utf-8 import boto3 import hcl import requests def get_terraform_vars(): s3_client = boto3.client("s3") tfvars_body = s3_client.get_object( Bucket="wellcomecollection-platform-infra", Key="terraform.tfvars" )["Body"] return hcl.load(tfvars_body) def build_url(es_credentials): protocol = es_credentials["protocol"] name = es_credentials["name"] region = es_credentials["region"] port = es_credentials["port"] return f"{protocol}://{name}.{region}.aws.found.io:{port}" def get_all_indexes(es_url, username, password): resp = requests.get( f"{es_url}/_cat/indices", auth=(username, password), params={"format": "json"} ) resp.raise_for_status() return resp.json() if __name__ == "__main__": terraform_vars = get_terraform_vars() es_cluster_credentials = terraform_vars["es_cluster_credentials"] es_url = build_url(es_cluster_credentials) username = es_cluster_credentials["username"] password = es_cluster_credentials["password"] indexes = get_all_indexes(es_url, username=username, password=password) print( '\n'.join(sorted( idx["index"] for idx in indexes if not idx["index"].startswith(".") )) )
Add a script for listing all Elasticsearch indexes
Add a script for listing all Elasticsearch indexes
Python
mit
wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api,wellcometrust/platform-api
--- +++ @@ -0,0 +1,54 @@ +#!/usr/bin/env python +# -*- encoding: utf-8 + +import boto3 +import hcl +import requests + + +def get_terraform_vars(): + s3_client = boto3.client("s3") + tfvars_body = s3_client.get_object( + Bucket="wellcomecollection-platform-infra", + Key="terraform.tfvars" + )["Body"] + return hcl.load(tfvars_body) + + +def build_url(es_credentials): + protocol = es_credentials["protocol"] + name = es_credentials["name"] + region = es_credentials["region"] + port = es_credentials["port"] + return f"{protocol}://{name}.{region}.aws.found.io:{port}" + + +def get_all_indexes(es_url, username, password): + resp = requests.get( + f"{es_url}/_cat/indices", + auth=(username, password), + params={"format": "json"} + ) + resp.raise_for_status() + + return resp.json() + + +if __name__ == "__main__": + terraform_vars = get_terraform_vars() + es_cluster_credentials = terraform_vars["es_cluster_credentials"] + + es_url = build_url(es_cluster_credentials) + + username = es_cluster_credentials["username"] + password = es_cluster_credentials["password"] + + indexes = get_all_indexes(es_url, username=username, password=password) + + print( + '\n'.join(sorted( + idx["index"] + for idx in indexes + if not idx["index"].startswith(".") + )) + )
006a921f19f6c4f64d694c86346ad85ada2c8bb8
tests/subclass_test.py
tests/subclass_test.py
#! /usr/bin/env python # -*- coding: utf-8 -*- # vi:ts=4:et try: import unittest2 as unittest except ImportError: import unittest import pycurl CLASSES = (pycurl.Curl, pycurl.CurlMulti, pycurl.CurlShare) class SubclassTest(unittest.TestCase): def test_baseclass_init(self): # base classes do not accept any arguments on initialization for baseclass in CLASSES: try: baseclass(0) except TypeError: pass else: raise AssertionError('Base class accepted invalid args') try: baseclass(a=1) except TypeError: pass else: raise AssertionError('Base class accepted invalid kwargs') def test_subclass_create(self): for baseclass in CLASSES: # test creation of a subclass class MyCurlClass(baseclass): pass # test creation of its object obj = MyCurlClass() # must be of type subclass, but also an instance of base class assert type(obj) == MyCurlClass assert isinstance(obj, baseclass) def test_subclass_init(self): for baseclass in CLASSES: class MyCurlClass(baseclass): def __init__(self, x, y=4): self.x = x self.y = y # subclass __init__ must be able to accept args and kwargs obj = MyCurlClass(3) assert obj.x == 3 assert obj.y == 4 obj = MyCurlClass(5, y=6) assert obj.x == 5 assert obj.y == 6 # and it must throw TypeError if arguments don't match try: MyCurlClass(1, 2, 3, kwarg=4) except TypeError: pass else: raise AssertionError('Subclass accepted invalid arguments') def test_subclass_method(self): for baseclass in CLASSES: class MyCurlClass(baseclass): def my_method(self, x): return x + 1 obj = MyCurlClass() # methods must be able to accept arguments and return a value assert obj.my_method(1) == 2 def test_subclass_method_override(self): # setopt args for each base class args = { pycurl.Curl: (pycurl.VERBOSE, 1), pycurl.CurlMulti: (pycurl.M_MAXCONNECTS, 3), pycurl.CurlShare: (pycurl.SH_SHARE, pycurl.LOCK_DATA_COOKIE), } for baseclass in CLASSES: class MyCurlClass(baseclass): def setopt(self, option, value): # base method must not be overwritten assert super().setopt != self.setopt # base method mut be callable, setopt must return None assert super().setopt(option, value) is None # return something else return 'my setopt' obj = MyCurlClass() assert obj.setopt(*args[baseclass]) == 'my setopt'
Add tests for subclass support
Add tests for subclass support
Python
lgpl-2.1
pycurl/pycurl,pycurl/pycurl,pycurl/pycurl
--- +++ @@ -0,0 +1,88 @@ +#! /usr/bin/env python +# -*- coding: utf-8 -*- +# vi:ts=4:et + +try: + import unittest2 as unittest +except ImportError: + import unittest +import pycurl + +CLASSES = (pycurl.Curl, pycurl.CurlMulti, pycurl.CurlShare) + +class SubclassTest(unittest.TestCase): + def test_baseclass_init(self): + # base classes do not accept any arguments on initialization + for baseclass in CLASSES: + try: + baseclass(0) + except TypeError: + pass + else: + raise AssertionError('Base class accepted invalid args') + try: + baseclass(a=1) + except TypeError: + pass + else: + raise AssertionError('Base class accepted invalid kwargs') + + def test_subclass_create(self): + for baseclass in CLASSES: + # test creation of a subclass + class MyCurlClass(baseclass): + pass + # test creation of its object + obj = MyCurlClass() + # must be of type subclass, but also an instance of base class + assert type(obj) == MyCurlClass + assert isinstance(obj, baseclass) + + def test_subclass_init(self): + for baseclass in CLASSES: + class MyCurlClass(baseclass): + def __init__(self, x, y=4): + self.x = x + self.y = y + # subclass __init__ must be able to accept args and kwargs + obj = MyCurlClass(3) + assert obj.x == 3 + assert obj.y == 4 + obj = MyCurlClass(5, y=6) + assert obj.x == 5 + assert obj.y == 6 + # and it must throw TypeError if arguments don't match + try: + MyCurlClass(1, 2, 3, kwarg=4) + except TypeError: + pass + else: + raise AssertionError('Subclass accepted invalid arguments') + + def test_subclass_method(self): + for baseclass in CLASSES: + class MyCurlClass(baseclass): + def my_method(self, x): + return x + 1 + obj = MyCurlClass() + # methods must be able to accept arguments and return a value + assert obj.my_method(1) == 2 + + def test_subclass_method_override(self): + # setopt args for each base class + args = { + pycurl.Curl: (pycurl.VERBOSE, 1), + pycurl.CurlMulti: (pycurl.M_MAXCONNECTS, 3), + pycurl.CurlShare: (pycurl.SH_SHARE, pycurl.LOCK_DATA_COOKIE), + } + for baseclass in CLASSES: + class MyCurlClass(baseclass): + def setopt(self, option, value): + # base method must not be overwritten + assert super().setopt != self.setopt + # base method mut be callable, setopt must return None + assert super().setopt(option, value) is None + # return something else + return 'my setopt' + obj = MyCurlClass() + assert obj.setopt(*args[baseclass]) == 'my setopt'
d764a483497afc5d029a82db14cc5cc88f45f4c0
nova/api/openstack/contrib/multinic.py
nova/api/openstack/contrib/multinic.py
# Copyright 2011 OpenStack LLC. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """The multinic extension.""" from webob import exc from nova import compute from nova import log as logging from nova.api.openstack import extensions from nova.api.openstack import faults LOG = logging.getLogger("nova.api.multinic") class Multinic(extensions.ExtensionDescriptor): def __init__(self, *args, **kwargs): super(Multinic, self).__init__(*args, **kwargs) self.compute_api = compute.API() def get_name(self): return "Multinic" def get_alias(self): return "NMN" def get_description(self): return "Multiple network support" def get_namespace(self): return "http://docs.openstack.org/ext/multinic/api/v1.1" def get_updated(self): return "2011-06-09T00:00:00+00:00" def get_actions(self): actions = [] # Add the add_fixed_ip action act = extensions.ActionExtension("servers", "addFixedIp", self._add_fixed_ip) actions.append(act) # Add the remove_fixed_ip action act = extensions.ActionExtension("servers", "removeFixedIp", self._remove_fixed_ip) actions.append(act) return actions def _add_fixed_ip(self, input_dict, req, id): """Adds an IP on a given network to an instance.""" try: # Validate the input entity if 'networkId' not in input_dict['addFixedIp']: LOG.exception(_("Missing 'networkId' argument for addFixedIp")) return faults.Fault(exc.HTTPUnprocessableEntity()) # Add the fixed IP network_id = input_dict['addFixedIp']['networkId'] self.compute_api.add_fixed_ip(req.environ['nova.context'], id, network_id) except Exception, e: LOG.exception(_("Error in addFixedIp %s"), e) return faults.Fault(exc.HTTPBadRequest()) return exc.HTTPAccepted() def _remove_fixed_ip(self, input_dict, req, id): # Not yet implemented raise faults.Fault(exc.HTTPNotImplemented())
Add an extension to allow for an addFixedIp action on instances
Add an extension to allow for an addFixedIp action on instances
Python
apache-2.0
saleemjaveds/https-github.com-openstack-nova,bigswitch/nova,virtualopensystems/nova,leilihh/nova,cyx1231st/nova,rajalokan/nova,whitepages/nova,eneabio/nova,devendermishrajio/nova,virtualopensystems/nova,luogangyi/bcec-nova,tanglei528/nova,cloudbau/nova,sileht/deb-openstack-nova,tianweizhang/nova,mandeepdhami/nova,salv-orlando/MyRepo,cloudbase/nova,devoid/nova,sridevikoushik31/openstack,JioCloud/nova_test_latest,savi-dev/nova,yosshy/nova,klmitch/nova,cernops/nova,dstroppa/openstack-smartos-nova-grizzly,psiwczak/openstack,BeyondTheClouds/nova,CiscoSystems/nova,kimjaejoong/nova,mikalstill/nova,SUSE-Cloud/nova,takeshineshiro/nova,hanlind/nova,redhat-openstack/nova,adelina-t/nova,viggates/nova,shahar-stratoscale/nova,gspilio/nova,maoy/zknova,maheshp/novatest,LoHChina/nova,DirectXMan12/nova-hacking,edulramirez/nova,rahulunair/nova,double12gzh/nova,MountainWei/nova,leilihh/nova,j-carpentier/nova,yrobla/nova,NewpTone/stacklab-nova,eneabio/nova,viggates/nova,imsplitbit/nova,zhimin711/nova,JianyuWang/nova,houshengbo/nova_vmware_compute_driver,vmturbo/nova,Triv90/Nova,gooddata/openstack-nova,JianyuWang/nova,double12gzh/nova,mmnelemane/nova,jianghuaw/nova,Francis-Liu/animated-broccoli,devendermishrajio/nova_test_latest,dawnpower/nova,usc-isi/extra-specs,JioCloud/nova_test_latest,savi-dev/nova,mikalstill/nova,rajalokan/nova,psiwczak/openstack,Yusuke1987/openstack_template,usc-isi/nova,rrader/nova-docker-plugin,citrix-openstack-build/nova,maoy/zknova,yrobla/nova,psiwczak/openstack,ntt-sic/nova,Yuriy-Leonov/nova,kimjaejoong/nova,rrader/nova-docker-plugin,cloudbase/nova,barnsnake351/nova,eayunstack/nova,tianweizhang/nova,orbitfp7/nova,phenoxim/nova,maelnor/nova,nikesh-mahalka/nova,orbitfp7/nova,maelnor/nova,tudorvio/nova,shahar-stratoscale/nova,cloudbau/nova,mahak/nova,gooddata/openstack-nova,joker946/nova,CCI-MOC/nova,mmnelemane/nova,zaina/nova,mgagne/nova,j-carpentier/nova,plumgrid/plumgrid-nova,gooddata/openstack-nova,openstack/nova,salv-orlando/MyRepo,adelina-t/nova,watonyweng/nova,berrange/nova,OpenAcademy-OpenStack/nova-scheduler,NewpTone/stacklab-nova,rajalokan/nova,belmiromoreira/nova,plumgrid/plumgrid-nova,LoHChina/nova,eonpatapon/nova,devoid/nova,usc-isi/nova,iuliat/nova,tanglei528/nova,leilihh/novaha,akash1808/nova_test_latest,barnsnake351/nova,JioCloud/nova,mikalstill/nova,fajoy/nova,edulramirez/nova,ewindisch/nova,russellb/nova,Yusuke1987/openstack_template,sacharya/nova,yatinkumbhare/openstack-nova,jeffrey4l/nova,ntt-sic/nova,isyippee/nova,sridevikoushik31/nova,NoBodyCam/TftpPxeBootBareMetal,ted-gould/nova,shootstar/novatest,bclau/nova,whitepages/nova,petrutlucian94/nova,Stavitsky/nova,josephsuh/extra-specs,fajoy/nova,sridevikoushik31/openstack,maheshp/novatest,jianghuaw/nova,belmiromoreira/nova,dims/nova,devendermishrajio/nova_test_latest,klmitch/nova,badock/nova,luogangyi/bcec-nova,josephsuh/extra-specs,openstack/nova,tangfeixiong/nova,mgagne/nova,mahak/nova,alvarolopez/nova,fajoy/nova,spring-week-topos/nova-week,hanlind/nova,saleemjaveds/https-github.com-openstack-nova,JioCloud/nova,cloudbase/nova-virtualbox,KarimAllah/nova,qwefi/nova,redhat-openstack/nova,mahak/nova,berrange/nova,alexandrucoman/vbox-nova-driver,thomasem/nova,maheshp/novatest,NeCTAR-RC/nova,scripnichenko/nova,vladikr/nova_drafts,yosshy/nova,angdraug/nova,Metaswitch/calico-nova,hanlind/nova,zzicewind/nova,vmturbo/nova,gooddata/openstack-nova,blueboxgroup/nova,rahulunair/nova,sridevikoushik31/openstack,CCI-MOC/nova,MountainWei/nova,alaski/nova,savi-dev/nova,NoBodyCam/TftpPxeBootBareMetal,leilihh/novaha,paulmathews/nova,CiscoSystems/nova,BeyondTheClouds/nova,rickerc/nova_audit,raildo/nova,silenceli/nova,petrutlucian94/nova_dev,cyx1231st/nova,CloudServer/nova,russellb/nova,usc-isi/nova,salv-orlando/MyRepo,eonpatapon/nova,eharney/nova,apporc/nova,Triv90/Nova,iuliat/nova,cernops/nova,jianghuaw/nova,bclau/nova,eharney/nova,imsplitbit/nova,klmitch/nova,NoBodyCam/TftpPxeBootBareMetal,nikesh-mahalka/nova,petrutlucian94/nova,mandeepdhami/nova,felixma/nova,watonyweng/nova,DirectXMan12/nova-hacking,Triv90/Nova,eayunstack/nova,takeshineshiro/nova,felixma/nova,paulmathews/nova,jeffrey4l/nova,sebrandon1/nova,affo/nova,vmturbo/nova,usc-isi/extra-specs,qwefi/nova,alvarolopez/nova,apporc/nova,gspilio/nova,bgxavier/nova,zzicewind/nova,projectcalico/calico-nova,Metaswitch/calico-nova,alexandrucoman/vbox-nova-driver,blueboxgroup/nova,TieWei/nova,Tehsmash/nova,sileht/deb-openstack-nova,rajalokan/nova,affo/nova,Juniper/nova,bigswitch/nova,dstroppa/openstack-smartos-nova-grizzly,aristanetworks/arista-ovs-nova,aristanetworks/arista-ovs-nova,sebrandon1/nova,sridevikoushik31/nova,paulmathews/nova,eneabio/nova,TieWei/nova,Juniper/nova,yrobla/nova,NewpTone/stacklab-nova,sebrandon1/nova,CEG-FYP-OpenStack/scheduler,KarimAllah/nova,Tehsmash/nova,TwinkleChawla/nova,KarimAllah/nova,usc-isi/extra-specs,TwinkleChawla/nova,OpenAcademy-OpenStack/nova-scheduler,scripnichenko/nova,houshengbo/nova_vmware_compute_driver,maoy/zknova,ewindisch/nova,ruslanloman/nova,silenceli/nova,akash1808/nova,citrix-openstack-build/nova,sileht/deb-openstack-nova,sacharya/nova,houshengbo/nova_vmware_compute_driver,tudorvio/nova,gspilio/nova,projectcalico/calico-nova,zhimin711/nova,yatinkumbhare/openstack-nova,Brocade-OpenSource/OpenStack-DNRM-Nova,Yuriy-Leonov/nova,tangfeixiong/nova,CloudServer/nova,thomasem/nova,SUSE-Cloud/nova,rickerc/nova_audit,Juniper/nova,cloudbase/nova-virtualbox,badock/nova,aristanetworks/arista-ovs-nova,fnordahl/nova,phenoxim/nova,cernops/nova,spring-week-topos/nova-week,BeyondTheClouds/nova,shootstar/novatest,tealover/nova,alaski/nova,sridevikoushik31/nova,devendermishrajio/nova,russellb/nova,dstroppa/openstack-smartos-nova-grizzly,akash1808/nova,shail2810/nova,tealover/nova,raildo/nova,NeCTAR-RC/nova,fnordahl/nova,klmitch/nova,vladikr/nova_drafts,DirectXMan12/nova-hacking,cloudbase/nova,zaina/nova,rahulunair/nova,ted-gould/nova,jianghuaw/nova,josephsuh/extra-specs,sridevikoushik31/nova,petrutlucian94/nova_dev,isyippee/nova,noironetworks/nova,shail2810/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,Francis-Liu/animated-broccoli,ruslanloman/nova,dims/nova,bgxavier/nova,noironetworks/nova,varunarya10/nova_test_latest,dawnpower/nova,varunarya10/nova_test_latest,vmturbo/nova,Juniper/nova,Stavitsky/nova,angdraug/nova,openstack/nova,akash1808/nova_test_latest,CEG-FYP-OpenStack/scheduler,joker946/nova
--- +++ @@ -0,0 +1,83 @@ +# Copyright 2011 OpenStack LLC. +# All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +"""The multinic extension.""" + +from webob import exc + +from nova import compute +from nova import log as logging +from nova.api.openstack import extensions +from nova.api.openstack import faults + + +LOG = logging.getLogger("nova.api.multinic") + + +class Multinic(extensions.ExtensionDescriptor): + def __init__(self, *args, **kwargs): + super(Multinic, self).__init__(*args, **kwargs) + self.compute_api = compute.API() + + def get_name(self): + return "Multinic" + + def get_alias(self): + return "NMN" + + def get_description(self): + return "Multiple network support" + + def get_namespace(self): + return "http://docs.openstack.org/ext/multinic/api/v1.1" + + def get_updated(self): + return "2011-06-09T00:00:00+00:00" + + def get_actions(self): + actions = [] + + # Add the add_fixed_ip action + act = extensions.ActionExtension("servers", "addFixedIp", + self._add_fixed_ip) + actions.append(act) + + # Add the remove_fixed_ip action + act = extensions.ActionExtension("servers", "removeFixedIp", + self._remove_fixed_ip) + actions.append(act) + + return actions + + def _add_fixed_ip(self, input_dict, req, id): + """Adds an IP on a given network to an instance.""" + try: + # Validate the input entity + if 'networkId' not in input_dict['addFixedIp']: + LOG.exception(_("Missing 'networkId' argument for addFixedIp")) + return faults.Fault(exc.HTTPUnprocessableEntity()) + + # Add the fixed IP + network_id = input_dict['addFixedIp']['networkId'] + self.compute_api.add_fixed_ip(req.environ['nova.context'], id, + network_id) + except Exception, e: + LOG.exception(_("Error in addFixedIp %s"), e) + return faults.Fault(exc.HTTPBadRequest()) + return exc.HTTPAccepted() + + def _remove_fixed_ip(self, input_dict, req, id): + # Not yet implemented + raise faults.Fault(exc.HTTPNotImplemented())
fe479bf2a8ec547922c6643bbdf0ba768eb79c9d
ludo/simulator.py
ludo/simulator.py
#!/usr/bin/env python3 from game import Game print("Welcome to a game of ludo!") average_throw_counter = 0 min_throws_per_game = 10000000 max_throws_per_game = 0 NUM_GAMES = 100 for i in range(0, NUM_GAMES): game = Game() throw_counter = 0 while game.next_move(): throw_counter += 1 average_throw_counter += throw_counter if throw_counter < min_throws_per_game: min_throws_per_game = throw_counter if throw_counter > max_throws_per_game: max_throws_per_game = throw_counter print("Game:", i+1) print("Average throws:", average_throw_counter/NUM_GAMES) print("Min", min_throws_per_game) print("Max", max_throws_per_game)
Add script to simulate multiple games
Add script to simulate multiple games
Python
mit
risteon/ludo_python
--- +++ @@ -0,0 +1,31 @@ +#!/usr/bin/env python3 +from game import Game + +print("Welcome to a game of ludo!") + +average_throw_counter = 0 +min_throws_per_game = 10000000 +max_throws_per_game = 0 +NUM_GAMES = 100 + +for i in range(0, NUM_GAMES): + + game = Game() + throw_counter = 0 + + while game.next_move(): + throw_counter += 1 + + average_throw_counter += throw_counter + if throw_counter < min_throws_per_game: + min_throws_per_game = throw_counter + if throw_counter > max_throws_per_game: + max_throws_per_game = throw_counter + + print("Game:", i+1) + +print("Average throws:", average_throw_counter/NUM_GAMES) +print("Min", min_throws_per_game) +print("Max", max_throws_per_game) + +
c89cce1a47c1e379958d7cced624ec0317cd3407
examples/demo3.py
examples/demo3.py
import os import sys sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) import logging import threading import xmpp2 import time import select from xmpp2 import XML # non-blocking, poll example. USERNAME = 'yourusername' PASSWORD = 'yourpassword' SERVER = 'example.com' logging.basicConfig(level=logging.DEBUG) logging.getLogger('xmpp2.xml.handler').setLevel(logging.INFO) c = xmpp2.Client(SERVER, stream_log_level=xmpp2.LOG_NONE) c.connect() c.auth(USERNAME, password=PASSWORD) c.write(XML.presence.add(XML.priority.add(1))) c.setblocking(False) po = select.poll() po.register(c, select.POLLIN) while True: for fd, event in po.poll(): msg = c.gen.next() if msg: sys.stdout.write(msg.pretty_print() + '\n')
Add demo for non-blocking with poll().
Add demo for non-blocking with poll().
Python
mit
easies/xmpp2
--- +++ @@ -0,0 +1,32 @@ +import os +import sys +sys.path.insert(0, os.path.join(os.path.dirname(__file__), '..')) +import logging +import threading +import xmpp2 +import time +import select +from xmpp2 import XML + +# non-blocking, poll example. +USERNAME = 'yourusername' +PASSWORD = 'yourpassword' +SERVER = 'example.com' + +logging.basicConfig(level=logging.DEBUG) +logging.getLogger('xmpp2.xml.handler').setLevel(logging.INFO) + +c = xmpp2.Client(SERVER, stream_log_level=xmpp2.LOG_NONE) +c.connect() +c.auth(USERNAME, password=PASSWORD) +c.write(XML.presence.add(XML.priority.add(1))) +c.setblocking(False) + +po = select.poll() +po.register(c, select.POLLIN) + +while True: + for fd, event in po.poll(): + msg = c.gen.next() + if msg: + sys.stdout.write(msg.pretty_print() + '\n')
1e65555a08ff3ee1a06e92d9dd054abf3cfaf711
media_tree/migrations/0003_alter_tree_fields.py
media_tree/migrations/0003_alter_tree_fields.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('media_tree', '0002_mptt_to_treebeard'), ] operations = [ migrations.AlterField( model_name='filenode', name='depth', field=models.PositiveIntegerField(db_index=True), ), migrations.AlterField( model_name='filenode', name='lft', field=models.PositiveIntegerField(db_index=True), ), migrations.AlterField( model_name='filenode', name='rgt', field=models.PositiveIntegerField(db_index=True), ), migrations.AlterField( model_name='filenode', name='tree_id', field=models.PositiveIntegerField(db_index=True), ), ]
Add a migration to update to final tree fields
Add a migration to update to final tree fields
Python
bsd-3-clause
samluescher/django-media-tree,samluescher/django-media-tree,samluescher/django-media-tree
--- +++ @@ -0,0 +1,34 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import models, migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ('media_tree', '0002_mptt_to_treebeard'), + ] + + operations = [ + migrations.AlterField( + model_name='filenode', + name='depth', + field=models.PositiveIntegerField(db_index=True), + ), + migrations.AlterField( + model_name='filenode', + name='lft', + field=models.PositiveIntegerField(db_index=True), + ), + migrations.AlterField( + model_name='filenode', + name='rgt', + field=models.PositiveIntegerField(db_index=True), + ), + migrations.AlterField( + model_name='filenode', + name='tree_id', + field=models.PositiveIntegerField(db_index=True), + ), + ]