commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
e85d0b100bd907ddf82f4c5e132690000d8cb4a0
smart_open/__init__.py
smart_open/__init__.py
# -*- coding: utf-8 -*- # # Copyright (C) 2019 Radim Rehurek <me@radimrehurek.com> # # This code is distributed under the terms and conditions # from the MIT License (MIT). # """ Utilities for streaming to/from several file-like data storages: S3 / HDFS / local filesystem / compressed files, and many more, using a simple, Pythonic API. The streaming makes heavy use of generators and pipes, to avoid loading full file contents into memory, allowing work with arbitrarily large files. The main functions are: * `open()`, which opens the given file for reading/writing * `parse_uri()` * `s3_iter_bucket()`, which goes over all keys in an S3 bucket in parallel * `register_compressor()`, which registers callbacks for transparent compressor handling """ import logging from smart_open import version logger = logging.getLogger(__name__) if len(logger.handlers) == 0: logger.addHandler(logging.NullHandler()) from .smart_open_lib import open, parse_uri, smart_open, register_compressor from .s3 import iter_bucket as s3_iter_bucket __all__ = [ 'open', 'parse_uri', 'register_compressor', 's3_iter_bucket', 'smart_open', ] __version__ = version.__version__
# -*- coding: utf-8 -*- # # Copyright (C) 2019 Radim Rehurek <me@radimrehurek.com> # # This code is distributed under the terms and conditions # from the MIT License (MIT). # """ Utilities for streaming to/from several file-like data storages: S3 / HDFS / local filesystem / compressed files, and many more, using a simple, Pythonic API. The streaming makes heavy use of generators and pipes, to avoid loading full file contents into memory, allowing work with arbitrarily large files. The main functions are: * `open()`, which opens the given file for reading/writing * `parse_uri()` * `s3_iter_bucket()`, which goes over all keys in an S3 bucket in parallel * `register_compressor()`, which registers callbacks for transparent compressor handling """ import logging from smart_open import version from .smart_open_lib import open, parse_uri, smart_open, register_compressor from .s3 import iter_bucket as s3_iter_bucket __all__ = [ 'open', 'parse_uri', 'register_compressor', 's3_iter_bucket', 'smart_open', ] __version__ = version.__version__ logger = logging.getLogger(__name__) logger.addHandler(logging.NullHandler())
Revert "Configure logging handlers before submodule imports"
Revert "Configure logging handlers before submodule imports" This reverts commit d9ce6cc440019ecfc73f1c82e41da4e9ce02a234.
Python
mit
RaRe-Technologies/smart_open,RaRe-Technologies/smart_open,piskvorky/smart_open
--- +++ @@ -25,10 +25,6 @@ import logging from smart_open import version -logger = logging.getLogger(__name__) -if len(logger.handlers) == 0: - logger.addHandler(logging.NullHandler()) - from .smart_open_lib import open, parse_uri, smart_open, register_compressor from .s3 import iter_bucket as s3_iter_bucket @@ -40,4 +36,8 @@ 'smart_open', ] + __version__ = version.__version__ + +logger = logging.getLogger(__name__) +logger.addHandler(logging.NullHandler())
4bc34f0b8adfc22037475cbdafbb149b7ea88421
tests/test_utils.py
tests/test_utils.py
import sys from girder_worker.utils import TeeStdOutCustomWrite, TeeStdErrCustomWrite, JobManager def test_TeeStdOutCustomeWrite(capfd): _nonlocal = {'data': ''} def _append_to_data(message, **kwargs): _nonlocal['data'] += message with TeeStdOutCustomWrite(_append_to_data): sys.stdout.write('Test String') sys.stdout.flush() assert _nonlocal['data'] == 'Test String'
import sys from girder_worker.utils import TeeStdOutCustomWrite def test_TeeStdOutCustomWrite(capfd): nonlocal_ = {'data': ''} def _append_to_data(message, **kwargs): nonlocal_['data'] += message with TeeStdOutCustomWrite(_append_to_data): sys.stdout.write('Test String') sys.stdout.flush() assert nonlocal_['data'] == 'Test String' out, err = capfd.readouterr() assert out == 'Test String'
Test TeeStdOutCustomWrite writes through to stdout
Test TeeStdOutCustomWrite writes through to stdout
Python
apache-2.0
girder/girder_worker,girder/girder_worker,girder/girder_worker
--- +++ @@ -1,15 +1,18 @@ import sys -from girder_worker.utils import TeeStdOutCustomWrite, TeeStdErrCustomWrite, JobManager +from girder_worker.utils import TeeStdOutCustomWrite -def test_TeeStdOutCustomeWrite(capfd): - _nonlocal = {'data': ''} +def test_TeeStdOutCustomWrite(capfd): + nonlocal_ = {'data': ''} def _append_to_data(message, **kwargs): - _nonlocal['data'] += message + nonlocal_['data'] += message with TeeStdOutCustomWrite(_append_to_data): sys.stdout.write('Test String') sys.stdout.flush() - assert _nonlocal['data'] == 'Test String' + assert nonlocal_['data'] == 'Test String' + + out, err = capfd.readouterr() + assert out == 'Test String'
08116d65cad849a2b031f960ef5da29e7802f4b7
tests/unit_tests.py
tests/unit_tests.py
import os import sys import unittest import tempfile sys.path.insert(0, os.environ.get('BLOG_PATH')) from blog.blog import * class BlogUnitTestCase(unittest.TestCase): def test_connect_db(self): db = connect_db() assert isinstance(db, sqlite3.Connection) def test_get_db(self): self.db, app.config['DATABASE'] = tempfile.mkstemp() app.config['TESTING'] = True self.app = app.test_client() with app.app_context(): db = get_db() assert isinstance(db, sqlite3.Connection) def schema(): return db.execute("SELECT * FROM sqlite_master").fetchall() assert len(schema()) == 0 init = migrate_db() assert len(schema()) == 11 if __name__ == '__main__': unittest.main()
import os import sys import unittest import tempfile sys.path.insert(0, os.environ.get('BLOG_PATH')) from blog.blog import * class BlogUnitTestCase(unittest.TestCase): def test_connect_db(self): db = connect_db() assert isinstance(db, sqlite3.Connection) def test_get_db(self): self.db, app.config['DATABASE'] = tempfile.mkstemp() app.config['TESTING'] = True self.app = app.test_client() with app.app_context(): db = get_db() assert isinstance(db, sqlite3.Connection) def schema(): return db.execute("SELECT * FROM sqlite_master").fetchall() assert len(schema()) == 0 init = migrate_db() assert len(schema()) == 21 if __name__ == '__main__': unittest.main()
Update count of things in DB
Update count of things in DB
Python
agpl-3.0
levlaz/blog,levlaz/blog,levlaz/blog,levlaz/blog
--- +++ @@ -27,7 +27,7 @@ init = migrate_db() - assert len(schema()) == 11 + assert len(schema()) == 21 if __name__ == '__main__': unittest.main()
ff6502fd8ecc4ee28cb05cb7ee8f11f75240ce47
mini_project.py
mini_project.py
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Other stuff from cothread.catools import caget, caput, ca_nothing # Load the machine ap.machines.load('SRI21') # First task BPMS = ap.getElements('BPM') print('There are {} BPM elements in the machine.'.format(len(BPMS))) # Second task print('A list of all the PV names for all BPMS') for BPM in range(len(BPMS)): # print(BPMS[BPM].pv()) print caget(BPMS[BPM].pv()) # Third task QUADS = ap.getElements('QUAD') print('String values for the setpoint currents') #for QUAD in range(len(QUADS)): # print(caget(QUADS[QUAD].pv(handle='readback')))
# Load the machine import pkg_resources pkg_resources.require('aphla') import aphla as ap # Other stuff from cothread.catools import caget, caput, ca_nothing # Load the machine ap.machines.load('SRI21') # First task BPMS = ap.getElements('BPM') print('There are {} BPM elements in the machine.'.format(len(BPMS))) # Second task print('A list of all the PV names for all BPMS') for BPM in range(len(BPMS)): pvs = BPMS[BPM].pv() print('PV names: {} PV values: {}'.format(pvs, caget(pvs))) # Third task QUADS = ap.getElements('QUAD') print('String values for the setpoint currents') #for QUAD in range(len(QUADS)): # print(caget(QUADS[QUAD].pv(handle='readback')))
Make the difference between PV names and values clearer.
Make the difference between PV names and values clearer.
Python
apache-2.0
razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects,razvanvasile/Work-Mini-Projects
--- +++ @@ -15,8 +15,8 @@ # Second task print('A list of all the PV names for all BPMS') for BPM in range(len(BPMS)): - # print(BPMS[BPM].pv()) - print caget(BPMS[BPM].pv()) + pvs = BPMS[BPM].pv() + print('PV names: {} PV values: {}'.format(pvs, caget(pvs))) # Third task QUADS = ap.getElements('QUAD')
c1c66caa31506190944c8bd37df6fe056b60c0e2
vote.py
vote.py
import enki import json e = enki.Enki('key', 'http://localhost:5001', 'translations') e.get_all() tasks = [] for t in e.tasks: options = dict() i = 0 for k in e.task_runs_df[t.id]['msgid'].keys(): options[k] = e.task_runs_df[t.id]['msgid'][k] t.info['msgid_options'] = options tasks.append(t.info) file = open('/tmp/translations_voting_tasks.json', 'w') file.write(json.dumps(tasks)) file.close()
import enki import json e = enki.Enki('key', 'http://localhost:5001', 'translations') e.get_all() tasks = [] for t in e.tasks: options = [] i = 0 for k in e.task_runs_df[t.id]['msgid'].keys(): option = dict(task_run_id=None, msgid=None) option['task_run_id'] = k option['msgid'] = e.task_runs_df[t.id]['msgid'][k] options.append(option) t.info['msgid_options'] = options tasks.append(t.info) file = open('/tmp/translations_voting_tasks.json', 'w') file.write(json.dumps(tasks)) file.close()
Save the tasks in the proper way.
Save the tasks in the proper way.
Python
agpl-3.0
PyBossa/app-translations
--- +++ @@ -9,10 +9,13 @@ tasks = [] for t in e.tasks: - options = dict() + options = [] i = 0 for k in e.task_runs_df[t.id]['msgid'].keys(): - options[k] = e.task_runs_df[t.id]['msgid'][k] + option = dict(task_run_id=None, msgid=None) + option['task_run_id'] = k + option['msgid'] = e.task_runs_df[t.id]['msgid'][k] + options.append(option) t.info['msgid_options'] = options tasks.append(t.info)
01c5a8d0f7e6540995817fecfee124f1859fa448
mistraldashboard/executions/views.py
mistraldashboard/executions/views.py
# -*- coding: utf-8 -*- # # Copyright 2014 - StackStorm, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from horizon import tables from mistraldashboard import api from mistraldashboard.executions.tables import ExecutionsTable from mistraldashboard.executions.tables import TaskTable class IndexView(tables.DataTableView): table_class = ExecutionsTable template_name = 'mistral/executions/index.html' def get_data(self): return api.execution_list(self.request) class TaskView(tables.DataTableView): table_class = TaskTable template_name = 'mistral/executions/index.html' def get_data(self): return api.task_list(self.kwargs['execution_id'])
# -*- coding: utf-8 -*- # # Copyright 2014 - StackStorm, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from horizon import tables from mistraldashboard import api from mistraldashboard.executions.tables import ExecutionsTable from mistraldashboard.executions.tables import TaskTable class IndexView(tables.DataTableView): table_class = ExecutionsTable template_name = 'mistral/executions/index.html' def get_data(self): return api.execution_list(self.request) class TaskView(tables.DataTableView): table_class = TaskTable template_name = 'mistral/executions/index.html' def get_data(self): return api.task_list(self.request, self.kwargs['execution_id'])
Fix errors when user click execution id
Fix errors when user click execution id Missing request argument when calling task list Change-Id: Idc282c267fc99513dd96a80f6f4a745bfefec2c8 Closes-Bug: #1471778
Python
apache-2.0
openstack/mistral-dashboard,openstack/mistral-dashboard,openstack/mistral-dashboard
--- +++ @@ -34,4 +34,4 @@ template_name = 'mistral/executions/index.html' def get_data(self): - return api.task_list(self.kwargs['execution_id']) + return api.task_list(self.request, self.kwargs['execution_id'])
10360cc4d956faac194c58eb3b52fae2b348b356
links/functions.py
links/functions.py
from uuid import uuid4 from random import randint from basic_http import BasicHttp def random_pk(length=8): s = uuid4().hex f = randint(0, (len(s) - length)) t = f + length return s[f:t] def get_url_info(url): print url req = BasicHttp(url) print req res = req.HEAD() data = { 'file_name': url.split('/')[-1], 'content_type': res['headers']['Content-Type'], 'file_size': res['headers']['Content-Length'] } return data
from uuid import uuid4 from random import randint from basic_http import BasicHttp def random_pk(length=8): s = uuid4().hex f = randint(0, (len(s) - length)) t = f + length return s[f:t] def get_url_info(url): req = BasicHttp(url) res = req.HEAD() data = { 'file_name': url.split('/')[-1], 'content_type': res['header']['Content-Type'], 'file_size': res['header']['Content-Length'] } return data
Remove of debug prints. Fix of invalid key name.
Remove of debug prints. Fix of invalid key name.
Python
bsd-3-clause
nachopro/followlink,nachopro/followlink
--- +++ @@ -11,15 +11,13 @@ return s[f:t] def get_url_info(url): - print url req = BasicHttp(url) - print req res = req.HEAD() data = { 'file_name': url.split('/')[-1], - 'content_type': res['headers']['Content-Type'], - 'file_size': res['headers']['Content-Length'] + 'content_type': res['header']['Content-Type'], + 'file_size': res['header']['Content-Length'] } return data
72384b3f06d4c68a94805e101f6cf4f820157834
process.py
process.py
# encoding: utf-8 import sys lines = sys.stdin.readlines() # Contact details are expected to begin on the fourth line, following the # header and a blank line, and extend until the next blank line. Lines with # bullets (•) will be split into separate lines. contact_lines = [] for line in lines[3:]: lines.remove(line) parts = [x.strip() for x in line.split("•")] if parts == ['']: break contact_lines.extend(parts) if '--tex' in sys.argv: lines.insert(0, "\\begin{nospace}\\begin{flushright}\n" + "\n\n".join(contact_lines) + "\n\\end{flushright}\\end{nospace}\n") print "".join(lines) if '--html' in sys.argv: lines.insert(0, "<div id='container'><div id='contact'>%s</div>\n" % "<br>".join(contact_lines)) lines.insert(1, "<div>") lines.append("</div>") print "".join(lines)
# encoding: utf-8 import sys lines = sys.stdin.readlines() # Contact details are expected to begin on the fourth line, following the # header and a blank line, and extend until the next blank line. Lines with # bullets (•) will be split into separate lines. contact_lines = [] for line in lines[3:]: lines.remove(line) parts = [x.strip() for x in line.split("•")] if parts == ['']: break contact_lines.extend(parts) if '--tex' in sys.argv: lines.insert(0, "\\begin{nospace}\\begin{flushright}\n" + "\n\n".join(contact_lines) + "\n\\end{flushright}\\end{nospace}\n") print "".join(lines).replace('~', '$\sim$') if '--html' in sys.argv: lines.insert(0, "<div id='container'><div id='contact'>%s</div>\n" % "<br>".join(contact_lines)) lines.insert(1, "<div>") lines.append("</div>") print "".join(lines)
Fix display of tildes in PDF output.
Fix display of tildes in PDF output.
Python
apache-2.0
davidbradway/resume,davidbradway/resume
--- +++ @@ -21,7 +21,7 @@ "\n\n".join(contact_lines) + "\n\\end{flushright}\\end{nospace}\n") - print "".join(lines) + print "".join(lines).replace('~', '$\sim$') if '--html' in sys.argv: lines.insert(0, "<div id='container'><div id='contact'>%s</div>\n" %
4581f05e937800b0541c219fd82390fdfef7644f
opal/tests/test_updates_from_dict.py
opal/tests/test_updates_from_dict.py
from django.test import TestCase from django.db import models as djangomodels from opal.models import UpdatesFromDictMixin class UpdatesFromDictMixinTest(TestCase): class TestDiagnosis(UpdatesFromDictMixin, djangomodels.Model): condition = djangomodels.CharField(max_length=255, blank=True, null=True) provisional = djangomodels.BooleanField() details = djangomodels.CharField(max_length=255, blank=True) date_of_diagnosis = djangomodels.DateField(blank=True, null=True) def test_get_fieldnames_to_serialise(self): names = self.TestDiagnosis._get_fieldnames_to_serialize() expected = ['id', 'condition', 'provisional', 'details', 'date_of_diagnosis'] self.assertEqual(expected, names)
from django.test import TestCase from django.db import models as djangomodels from opal.models import UpdatesFromDictMixin class UpdatesFromDictMixinTest(TestCase): class TestDiagnosis(UpdatesFromDictMixin, djangomodels.Model): condition = djangomodels.CharField(max_length=255, blank=True, null=True) provisional = djangomodels.BooleanField() details = djangomodels.CharField(max_length=255, blank=True) date_of_diagnosis = djangomodels.DateField(blank=True, null=True) def test_get_fieldnames_to_serialise(self): names = self.TestDiagnosis._get_fieldnames_to_serialize() expected = ['id', 'condition', 'provisional', 'details', 'date_of_diagnosis'] self.assertEqual(expected, names) def test_get_named_foreign_key_fields(self): for name in ['patient_id', 'episode_id', 'gp_id', 'nurse_id']: self.assertEqual(djangomodels.ForeignKey, self.TestDiagnosis._get_field_type(name))
Add a specific test for the updatesfromdict hard coded foreignkey fields
Add a specific test for the updatesfromdict hard coded foreignkey fields
Python
agpl-3.0
khchine5/opal,khchine5/opal,khchine5/opal
--- +++ @@ -15,3 +15,7 @@ expected = ['id', 'condition', 'provisional', 'details', 'date_of_diagnosis'] self.assertEqual(expected, names) + def test_get_named_foreign_key_fields(self): + for name in ['patient_id', 'episode_id', 'gp_id', 'nurse_id']: + self.assertEqual(djangomodels.ForeignKey, + self.TestDiagnosis._get_field_type(name))
3520217e38849ad18b11245c6cac51d79db8422d
pytablereader/loadermanager/_base.py
pytablereader/loadermanager/_base.py
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import from ..interface import TableLoaderInterface class TableLoaderManager(TableLoaderInterface): def __init__(self, loader): self.__loader = loader @property def loader(self): return self.__loader @property def format_name(self): return self.__loader.format_name @property def source_type(self): return self.__loader.source_type @property def encoding(self): try: return self.__loader.encoding except AttributeError: return None @encoding.setter def encoding(self, codec_name): self.__loader.encoding = codec_name def load(self): return self.__loader.load() def inc_table_count(self): self.__loader.inc_table_count()
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import from ..interface import TableLoaderInterface class TableLoaderManager(TableLoaderInterface): def __init__(self, loader): self.__loader = loader @property def loader(self): return self.__loader @property def format_name(self): return self.__loader.format_name @property def source_type(self): return self.__loader.source_type @property def table_name(self): return self.__loader.table_name @table_name.setter def table_name(self, value): self.__loader.table_name = value @property def encoding(self): try: return self.__loader.encoding except AttributeError: return None @encoding.setter def encoding(self, codec_name): self.__loader.encoding = codec_name def load(self): return self.__loader.load() def inc_table_count(self): self.__loader.inc_table_count()
Add an interface to change table_name
Add an interface to change table_name
Python
mit
thombashi/pytablereader,thombashi/pytablereader,thombashi/pytablereader
--- +++ @@ -27,6 +27,14 @@ return self.__loader.source_type @property + def table_name(self): + return self.__loader.table_name + + @table_name.setter + def table_name(self, value): + self.__loader.table_name = value + + @property def encoding(self): try: return self.__loader.encoding
ec1058fc1576b7a47af2d00b5b9e60e71dc6dc6b
kraller/kraller.py
kraller/kraller.py
""" Kraller An application to allow signups for accounts on a server with a key. """ from flask import Flask app = Flask(__name__) @app.route('/') def index(): return 'Hello from ACM@CWRU' if __name__ == '__main__': app.run()
""" Kraller An application to allow signups for accounts on a server with a key. """ from flask import Flask app = Flask(__name__) @app.route('/') def index(): return 'Hello from ACM@CWRU' """ Don't invoke this directly in production. This is for development only. Use a WSGI webserver such as Gunicorn to serve this in production. """ if __name__ == '__main__': app.run(debug=True)
Add debug mode to the development version
Add debug mode to the development version
Python
bsd-2-clause
Aaronneyer/kraller,Aaronneyer/kraller,bis12/kraller,hashbrowncipher/kraller,hacsoc/kraller,hashbrowncipher/kraller,hacsoc/kraller,cwruacm/kraller,bis12/kraller,cwruacm/kraller
--- +++ @@ -14,5 +14,9 @@ return 'Hello from ACM@CWRU' +""" +Don't invoke this directly in production. This is for development only. +Use a WSGI webserver such as Gunicorn to serve this in production. +""" if __name__ == '__main__': - app.run() + app.run(debug=True)
d518c4f8fa7b657c735d5a2b4f653d5c0cad529e
gnotty/templatetags/gnotty_tags.py
gnotty/templatetags/gnotty_tags.py
from django import template from django.db.models import Min, Max from gnotty.models import IRCMessage from gnotty.conf import settings register = template.Library() @register.inclusion_tag("gnotty/includes/nav.html", takes_context=True) def gnotty_nav(context): min_max = IRCMessage.objects.aggregate(Min("message_time"), Max("message_time")) if min_max: years = range(min_max["message_time__max"].year, min_max["message_time__min"].year - 1, -1) else: years = [] context["IRC_CHANNEL"] = settings.IRC_CHANNEL context["years"] = years return context
from django import template from django.db.models import Min, Max from gnotty.models import IRCMessage from gnotty.conf import settings register = template.Library() @register.inclusion_tag("gnotty/includes/nav.html", takes_context=True) def gnotty_nav(context): min_max = IRCMessage.objects.aggregate(Min("message_time"), Max("message_time")) if min_max.values()[0]: years = range(min_max["message_time__max"].year, min_max["message_time__min"].year - 1, -1) else: years = [] context["IRC_CHANNEL"] = settings.IRC_CHANNEL context["years"] = years return context
Fix check for empty min/max years.
Fix check for empty min/max years.
Python
bsd-2-clause
spaceone/gnotty,stephenmcd/gnotty,stephenmcd/gnotty,spaceone/gnotty,stephenmcd/gnotty,spaceone/gnotty
--- +++ @@ -13,7 +13,7 @@ def gnotty_nav(context): min_max = IRCMessage.objects.aggregate(Min("message_time"), Max("message_time")) - if min_max: + if min_max.values()[0]: years = range(min_max["message_time__max"].year, min_max["message_time__min"].year - 1, -1) else:
087b6a623e4a48b76fa3ce62a14298d9744afd2a
go/apps/bulk_message/definition.py
go/apps/bulk_message/definition.py
from go.vumitools.conversation.definition import ( ConversationDefinitionBase, ConversationAction) class BulkSendAction(ConversationAction): action_name = 'bulk_send' action_display_name = 'Write and send bulk message' action_display_verb = 'Send message' needs_confirmation = True needs_group = True needs_running = True def check_disabled(self): if self._conv.has_channel_supporting_generic_sends(): return None return ("This action needs channels capable of sending" " messages attached to this conversation.") def perform_action(self, action_data): return self.send_command( 'bulk_send', batch_id=self._conv.batch.key, msg_options={}, content=action_data['message'], delivery_class=action_data['delivery_class'], dedupe=action_data['dedupe']) class ConversationDefinition(ConversationDefinitionBase): conversation_type = 'bulk_message' actions = (BulkSendAction,)
from go.scheduler.models import Task from go.vumitools.conversation.definition import ( ConversationDefinitionBase, ConversationAction) class BulkSendAction(ConversationAction): action_name = 'bulk_send' action_display_name = 'Write and send bulk message' action_display_verb = 'Send message now' action_schedule_verb = 'Schedule' needs_confirmation = True needs_group = True needs_running = True def check_disabled(self): if self._conv.has_channel_supporting_generic_sends(): return None return ("This action needs channels capable of sending" " messages attached to this conversation.") def perform_action(self, action_data): if action_data['scheduled_datetime'] is None: return self.send_command( 'bulk_send', batch_id=self._conv.batch.key, msg_options={}, content=action_data['message'], delivery_class=action_data['delivery_class'], dedupe=action_data['dedupe']) else: task = Task.objects.create( account_id=self._conv.user_api.user_account_key, label='Bulk Message Send', task_type=Task.TYPE_CONVERSATION_ACTION, task_data={ 'action_name': 'bulk_send', 'action_kwargs': { 'batch_id': self._conv.batch.key, 'msg_options': {}, 'content': action_data['message'], 'delivery_class': action_data['delivery_class'], 'dedupe': action_data['dedupe'], }, }, scheduled_for=action_data['scheduled_datetime']) task.save() class ConversationDefinition(ConversationDefinitionBase): conversation_type = 'bulk_message' actions = (BulkSendAction,)
Add handling for creating scheduled action for bulk send
Add handling for creating scheduled action for bulk send
Python
bsd-3-clause
praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go,praekelt/vumi-go
--- +++ @@ -1,3 +1,4 @@ +from go.scheduler.models import Task from go.vumitools.conversation.definition import ( ConversationDefinitionBase, ConversationAction) @@ -5,7 +6,8 @@ class BulkSendAction(ConversationAction): action_name = 'bulk_send' action_display_name = 'Write and send bulk message' - action_display_verb = 'Send message' + action_display_verb = 'Send message now' + action_schedule_verb = 'Schedule' needs_confirmation = True @@ -19,11 +21,29 @@ " messages attached to this conversation.") def perform_action(self, action_data): - return self.send_command( - 'bulk_send', batch_id=self._conv.batch.key, - msg_options={}, content=action_data['message'], - delivery_class=action_data['delivery_class'], - dedupe=action_data['dedupe']) + if action_data['scheduled_datetime'] is None: + return self.send_command( + 'bulk_send', batch_id=self._conv.batch.key, + msg_options={}, content=action_data['message'], + delivery_class=action_data['delivery_class'], + dedupe=action_data['dedupe']) + else: + task = Task.objects.create( + account_id=self._conv.user_api.user_account_key, + label='Bulk Message Send', + task_type=Task.TYPE_CONVERSATION_ACTION, + task_data={ + 'action_name': 'bulk_send', + 'action_kwargs': { + 'batch_id': self._conv.batch.key, + 'msg_options': {}, + 'content': action_data['message'], + 'delivery_class': action_data['delivery_class'], + 'dedupe': action_data['dedupe'], + }, + }, + scheduled_for=action_data['scheduled_datetime']) + task.save() class ConversationDefinition(ConversationDefinitionBase):
c38fb74a71b3471f5363cf1aa95fecdd8ac9180f
picoCTF-shell/tests/test_problems.py
picoCTF-shell/tests/test_problems.py
import pytest from os.path import dirname, join, realpath import hacksport.deploy from hacksport.deploy import deploy_problem from shell_manager.util import default_config PATH = dirname(realpath(__file__)) hacksport.deploy.deploy_config = default_config @pytest.mark.skip("Broken tests/not working") class TestProblems: """ Regression tests for compiled problems. """ def test_compiled_sources(self): deploy_problem(join(PATH, "problems/compiled_sources"), test=True) def test_remote_compiled_makefile_template(self): deploy_problem( join(PATH, "problems/remote_compiled_makefile_template"), test=True) def test_remote_no_compile(self): deploy_problem(join(PATH, "problems/remote_no_compile"), test=True) def test_compiled_sources_url(self): deploy_problem(join(PATH, "problems/compiled_sources_url"), test=True) def test_high_level_compiled_binary(self): deploy_problem(join(PATH, "problems/local_compiled1"), test=True) deploy_problem(join(PATH, "problems/local_compiled2"), test=True) deploy_problem(join(PATH, "problems/remote_compiled1"), test=True)
import pytest from os.path import dirname, join, realpath import hacksport.deploy from hacksport.deploy import deploy_problem PATH = dirname(realpath(__file__)) @pytest.mark.skip("Broken tests/not working") class TestProblems: """ Regression tests for compiled problems. """ def test_compiled_sources(self): deploy_problem(join(PATH, "problems/compiled_sources"), test=True) def test_remote_compiled_makefile_template(self): deploy_problem( join(PATH, "problems/remote_compiled_makefile_template"), test=True) def test_remote_no_compile(self): deploy_problem(join(PATH, "problems/remote_no_compile"), test=True) def test_compiled_sources_url(self): deploy_problem(join(PATH, "problems/compiled_sources_url"), test=True) def test_high_level_compiled_binary(self): deploy_problem(join(PATH, "problems/local_compiled1"), test=True) deploy_problem(join(PATH, "problems/local_compiled2"), test=True) deploy_problem(join(PATH, "problems/remote_compiled1"), test=True)
Remove broken import in skipped tests
Remove broken import in skipped tests
Python
mit
royragsdale/picoCTF,picoCTF/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,royragsdale/picoCTF,royragsdale/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF,royragsdale/picoCTF,picoCTF/picoCTF
--- +++ @@ -3,11 +3,9 @@ import hacksport.deploy from hacksport.deploy import deploy_problem -from shell_manager.util import default_config PATH = dirname(realpath(__file__)) -hacksport.deploy.deploy_config = default_config @pytest.mark.skip("Broken tests/not working")
36f218ccab5de6799674c438546189245b444b25
was/photo/forms.py
was/photo/forms.py
from django.forms.models import ModelForm from .models import Photo from artists.models import Artists class UploadPhotoForm(ModelForm): class Meta: model = Photo fields = ['picture', 'comment'] def __init__(self, *args, **kwargs): self.request = kwargs.pop("request") super(UploadPhotoForm, self).__init__(*args, **kwargs) def save(self): photo = super(UploadPhotoForm, self).save(commit=False) artist = Artists.objects.get(user=self.request.user) photo.artist = artist photo.save() return photo
from django.forms.models import ModelForm from .models import Photo from artists.models import Artists class UploadPhotoForm(ModelForm): class Meta: model = Photo fields = ['picture', 'comment'] def __init__(self, *args, **kwargs): self.request = kwargs.pop("request", None) super(UploadPhotoForm, self).__init__(*args, **kwargs) def save(self): photo = super(UploadPhotoForm, self).save(commit=False) artist = Artists.objects.get(user=self.request.user) photo.artist = artist photo.save() return photo
Add None when request arg is pop, in case we access the form in get
Add None when request arg is pop, in case we access the form in get
Python
mit
KeserOner/where-artists-share,KeserOner/where-artists-share
--- +++ @@ -9,7 +9,7 @@ fields = ['picture', 'comment'] def __init__(self, *args, **kwargs): - self.request = kwargs.pop("request") + self.request = kwargs.pop("request", None) super(UploadPhotoForm, self).__init__(*args, **kwargs) def save(self):
224478e0d69197feb19f2f89c21e884a2e5cb77d
organizer/views.py
organizer/views.py
from django.shortcuts import ( get_object_or_404, render) from .models import Startup, Tag def startup_detail(request, slug): startup = get_object_or_404( Startup, slug__iexact=slug) return render( request, 'organizer/startup_detail.html', {'startup': startup}) def startup_list(request): return render( request, 'organizer/startup_list.html', {'startup_list': Startup.objects.all()}) def tag_create(request): pass def tag_detail(request, slug): tag = get_object_or_404( Tag, slug__iexact=slug) return render( request, 'organizer/tag_detail.html', {'tag': tag}) def tag_list(request): return render( request, 'organizer/tag_list.html', {'tag_list': Tag.objects.all()})
from django.shortcuts import ( get_object_or_404, render) from .models import Startup, Tag def startup_detail(request, slug): startup = get_object_or_404( Startup, slug__iexact=slug) return render( request, 'organizer/startup_detail.html', {'startup': startup}) def startup_list(request): return render( request, 'organizer/startup_list.html', {'startup_list': Startup.objects.all()}) def tag_create(request): if request.method == 'POST': # bind data to form # if the data is valid: # create new object from data # show webpage for new object # else: (empty data or invalid data) # show bound HTML form (with errors) pass else: # request.method != 'POST' # show unbound HTML form pass def tag_detail(request, slug): tag = get_object_or_404( Tag, slug__iexact=slug) return render( request, 'organizer/tag_detail.html', {'tag': tag}) def tag_list(request): return render( request, 'organizer/tag_list.html', {'tag_list': Tag.objects.all()})
Add HTTP method condition to tag_create().
Ch09: Add HTTP method condition to tag_create().
Python
bsd-2-clause
jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8
--- +++ @@ -21,7 +21,17 @@ def tag_create(request): - pass + if request.method == 'POST': + # bind data to form + # if the data is valid: + # create new object from data + # show webpage for new object + # else: (empty data or invalid data) + # show bound HTML form (with errors) + pass + else: # request.method != 'POST' + # show unbound HTML form + pass def tag_detail(request, slug):
271c91a607797bbd9d1a1e179745aedf580e8209
user/models.py
user/models.py
from django.conf import settings from django.contrib.auth.models import ( AbstractBaseUser, PermissionsMixin) from django.core.urlresolvers import reverse from django.db import models class Profile(models.Model): user = models.OneToOneField( settings.AUTH_USER_MODEL) slug = models.SlugField( max_length=30, unique=True) about = models.TextField() def __str__(self): return self.user.get_username() def get_absolute_url(self): return reverse( 'dj-auth:public_profile', kwargs={'slug': self.slug}) def get_update_url(self): return reverse('dj-auth:profile_update') class User(AbstractBaseUser, PermissionsMixin): email = models.EmailField( 'email address', max_length=254, unique=True) is_staff = models.BooleanField( 'staff status', default=False, help_text=( 'Designates whether the user can ' 'log into this admin site.')) is_active = models.BooleanField( 'active', default=True, help_text=( 'Designates whether this user should ' 'be treated as active. Unselect this ' 'instead of deleting accounts.')) USERNAME_FIELD = 'email' def __str__(self): return self.email def get_absolute_url(self): return self.profile.get_absolute_url()
from django.conf import settings from django.contrib.auth.models import ( AbstractBaseUser, PermissionsMixin) from django.core.urlresolvers import reverse from django.db import models class Profile(models.Model): user = models.OneToOneField( settings.AUTH_USER_MODEL) name = models.CharField( max_length=255) slug = models.SlugField( max_length=30, unique=True) about = models.TextField() joined = models.DateTimeField( "Date Joined", auto_now_add=True) def __str__(self): return self.user.get_username() def get_absolute_url(self): return reverse( 'dj-auth:public_profile', kwargs={'slug': self.slug}) def get_update_url(self): return reverse('dj-auth:profile_update') class User(AbstractBaseUser, PermissionsMixin): email = models.EmailField( 'email address', max_length=254, unique=True) is_staff = models.BooleanField( 'staff status', default=False, help_text=( 'Designates whether the user can ' 'log into this admin site.')) is_active = models.BooleanField( 'active', default=True, help_text=( 'Designates whether this user should ' 'be treated as active. Unselect this ' 'instead of deleting accounts.')) USERNAME_FIELD = 'email' def __str__(self): return self.email def get_absolute_url(self): return self.profile.get_absolute_url() def get_full_name(self): return self.profile.name def get_short_name(self): return self.profile.name
Add name and joined date field to Profile.
Ch22: Add name and joined date field to Profile.
Python
bsd-2-clause
jambonrose/DjangoUnleashed-1.8,jambonrose/DjangoUnleashed-1.8
--- +++ @@ -8,10 +8,15 @@ class Profile(models.Model): user = models.OneToOneField( settings.AUTH_USER_MODEL) + name = models.CharField( + max_length=255) slug = models.SlugField( max_length=30, unique=True) about = models.TextField() + joined = models.DateTimeField( + "Date Joined", + auto_now_add=True) def __str__(self): return self.user.get_username() @@ -51,3 +56,9 @@ def get_absolute_url(self): return self.profile.get_absolute_url() + + def get_full_name(self): + return self.profile.name + + def get_short_name(self): + return self.profile.name
1b1a40646a08e6a01927279eb7dff8399adec740
markups/common.py
markups/common.py
# This file is part of python-markups module # License: BSD # Copyright: (C) Dmitry Shachnev, 2012-2017 import os.path # Some common constants and functions (LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3) CONFIGURATION_DIR = (os.getenv('XDG_CONFIG_HOME') or os.getenv('APPDATA') or os.path.expanduser('~/.config')) MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' MATHJAX_WEB_URL = 'https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.1/MathJax.js' PYGMENTS_STYLE = 'default' def get_pygments_stylesheet(selector, style=None): if style is None: style = PYGMENTS_STYLE if style == '': return '' try: from pygments.formatters import HtmlFormatter except ImportError: return '' else: return HtmlFormatter(style=style).get_style_defs(selector) + '\n' def get_mathjax_url(webenv): if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv: return MATHJAX_LOCAL_URL else: return MATHJAX_WEB_URL
# This file is part of python-markups module # License: BSD # Copyright: (C) Dmitry Shachnev, 2012-2017 import os.path # Some common constants and functions (LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3) CONFIGURATION_DIR = (os.getenv('XDG_CONFIG_HOME') or os.getenv('APPDATA') or os.path.expanduser('~/.config')) MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' MATHJAX_WEB_URL = 'https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/MathJax.js' PYGMENTS_STYLE = 'default' def get_pygments_stylesheet(selector, style=None): if style is None: style = PYGMENTS_STYLE if style == '': return '' try: from pygments.formatters import HtmlFormatter except ImportError: return '' else: return HtmlFormatter(style=style).get_style_defs(selector) + '\n' def get_mathjax_url(webenv): if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv: return MATHJAX_LOCAL_URL else: return MATHJAX_WEB_URL
Update MathJax URL to 2.7.2
Update MathJax URL to 2.7.2
Python
bsd-3-clause
mitya57/pymarkups,retext-project/pymarkups
--- +++ @@ -9,7 +9,7 @@ CONFIGURATION_DIR = (os.getenv('XDG_CONFIG_HOME') or os.getenv('APPDATA') or os.path.expanduser('~/.config')) MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' -MATHJAX_WEB_URL = 'https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.1/MathJax.js' +MATHJAX_WEB_URL = 'https://cdnjs.cloudflare.com/ajax/libs/mathjax/2.7.2/MathJax.js' PYGMENTS_STYLE = 'default'
5a05b1cbe258ef18607aabbc276aaccd5f2eb14c
admin_crud/router.py
admin_crud/router.py
from django.apps import apps as dj_apps from django.conf.urls import include, url from django.template.response import TemplateResponse from django.utils.text import capfirst class Router(object): def __init__(self): self.registry = [] self._urls = [ url(r'^$', self.index_view, name='index') ] def build_groups(self, request): apps = self.registry models = {} for path, controller in apps: app_label = controller.model._meta.app_label if app_label not in models: models[app_label] = { 'verbose_name': dj_apps.get_app_config(app_label).verbose_name, 'admins': [] } controller_info = { 'verbose_name': capfirst(controller.model._meta.verbose_name_plural), } models[app_label]['admins'].append(controller_info) return models def index_view(self, request, *args, **kwargs): groups = self.build_groups(request) return TemplateResponse( request, template='admin_crud/index.html', context={ 'groups': groups } ) def register(self, path, Controller): controller = Controller() self.registry.append([path, controller]) self._urls += [ url(r'^%s/' % path, include(controller.get_urls())) ] @property def urls(self): return [self._urls, 'admin-crud', 'admin-crud']
from django.apps import apps as dj_apps from django.conf.urls import include, url from django.template.response import TemplateResponse from django.utils.text import capfirst class Router(object): def __init__(self): self.registry = [] self._urls = [ url(r'^$', self.index_view, name='index') ] def build_groups(self, request): apps = self.registry groups = {} for path, controller in apps: app_label = controller.model._meta.app_label if app_label not in groups: groups[app_label] = { 'verbose_name': dj_apps.get_app_config(app_label).verbose_name, 'admins': [] } controller_info = { 'verbose_name': capfirst(controller.model._meta.verbose_name_plural), } groups[app_label]['admins'].append(controller_info) return groups def index_view(self, request, *args, **kwargs): groups = self.build_groups(request) return TemplateResponse( request, template='admin_crud/index.html', context={ 'groups': groups } ) def register(self, path, Controller): controller = Controller() self.registry.append([path, controller]) self._urls += [ url(r'^%s/' % path, include(controller.get_urls())) ] @property def urls(self): return [self._urls, 'admin-crud', 'admin-crud']
Rename models to groups for corresponding with method name
Rename models to groups for corresponding with method name
Python
mit
raizanshine/django-admin-crud,raizanshine/django-admin-crud
--- +++ @@ -13,20 +13,21 @@ def build_groups(self, request): apps = self.registry - models = {} + groups = {} for path, controller in apps: app_label = controller.model._meta.app_label - if app_label not in models: - models[app_label] = { + if app_label not in groups: + groups[app_label] = { 'verbose_name': dj_apps.get_app_config(app_label).verbose_name, 'admins': [] } controller_info = { 'verbose_name': capfirst(controller.model._meta.verbose_name_plural), } - models[app_label]['admins'].append(controller_info) - return models + groups[app_label]['admins'].append(controller_info) + + return groups def index_view(self, request, *args, **kwargs): groups = self.build_groups(request)
e0233b0bb0a5a92c641d06648ec9921318b9a028
knowledge_repo/converters/pdf.py
knowledge_repo/converters/pdf.py
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'w') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf()
from ..converter import KnowledgePostConverter from .html import HTMLConverter class PDFConverter(KnowledgePostConverter): ''' Use this as a template for new KnowledgePostConverters. ''' _registry_keys = ['pdf'] @property def dependencies(self): # Dependencies required for this converter on top of core knowledge-repo dependencies return ['weasyprint'] def from_file(self, filename, **opts): raise NotImplementedError def from_string(self, filename, **opts): raise NotImplementedError def to_file(self, filename, **opts): with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts): from weasyprint import HTML html = HTMLConverter(self.kp).to_string() return HTML(string=html).write_pdf()
Fix encoding error in Python 3.
Fix encoding error in Python 3.
Python
apache-2.0
airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo,airbnb/knowledge-repo
--- +++ @@ -21,7 +21,7 @@ raise NotImplementedError def to_file(self, filename, **opts): - with open(filename, 'w') as f: + with open(filename, 'wb') as f: f.write(self.to_string()) def to_string(self, **opts):
4c43e7588d8747e26901400ce67c1485ba98bb92
l10n_br_purchase/__manifest__.py
l10n_br_purchase/__manifest__.py
# -*- coding: utf-8 -*- # Copyright (C) 2009 Renato Lima - Akretion # License AGPL-3 - See http://www.gnu.org/licenses/agpl-3.0.html { 'name': 'Brazilian Localization Purchase', 'license': 'AGPL-3', 'category': 'Localisation', 'author': 'Akretion, Odoo Community Association (OCA)', 'website': 'http://odoo-brasil.org', 'version': '8.0.1.0.0', 'depends': [ 'l10n_br_stock_account', 'account_fiscal_position_rule_purchase', ], 'data': [ 'data/l10n_br_purchase_data.xml', 'views/purchase_view.xml', 'views/res_company_view.xml', 'security/ir.model.access.csv', 'security/l10n_br_purchase_security.xml', ], 'demo': [ # FIXME # 'test/purchase_order_demo.yml' ], 'installable': False, 'auto_install': False, }
# -*- coding: utf-8 -*- # Copyright (C) 2009 Renato Lima - Akretion # License AGPL-3 - See http://www.gnu.org/licenses/agpl-3.0.html { 'name': 'Brazilian Localization Purchase', 'license': 'AGPL-3', 'category': 'Localisation', 'author': 'Akretion, Odoo Community Association (OCA)', 'website': 'http://odoo-brasil.org', 'version': '10.0.1.0.0', 'depends': [ 'l10n_br_stock_account', 'account_fiscal_position_rule_purchase', ], 'data': [ 'data/l10n_br_purchase_data.xml', 'views/purchase_view.xml', 'views/res_company_view.xml', 'security/ir.model.access.csv', 'security/l10n_br_purchase_security.xml', ], 'demo': [ # FIXME # 'test/purchase_order_demo.yml' ], 'installable': False, 'auto_install': False, }
Bump module version to 10.0.1.0.0
[MIG] Bump module version to 10.0.1.0.0
Python
agpl-3.0
OCA/l10n-brazil,akretion/l10n-brazil,akretion/l10n-brazil,OCA/l10n-brazil,OCA/l10n-brazil,akretion/l10n-brazil
--- +++ @@ -8,7 +8,7 @@ 'category': 'Localisation', 'author': 'Akretion, Odoo Community Association (OCA)', 'website': 'http://odoo-brasil.org', - 'version': '8.0.1.0.0', + 'version': '10.0.1.0.0', 'depends': [ 'l10n_br_stock_account', 'account_fiscal_position_rule_purchase',
6caca3259f4ec8f298b1d35f15e4492efbcff6b1
tests/basics/dict1.py
tests/basics/dict1.py
# basic dictionary d = {} print(d) d[2] = 123 print(d) d = {1:2} d[3] = 3 print(len(d), d[1], d[3]) d[1] = 0 print(len(d), d[1], d[3]) print(str(d) == '{1: 0, 3: 3}' or str(d) == '{3: 3, 1: 0}') x = 1 while x < 100: d[x] = x x += 1 print(d[50]) # equality operator on dicts of different size print({} == {1:1}) # equality operator on dicts of same size but with different keys print({1:1} == {2:1}) # value not found try: {}[0] except KeyError: print('KeyError') # unsupported unary op try: +{} except TypeError: print('TypeError') # unsupported binary op try: {} + {} except TypeError: print('TypeError')
# basic dictionary d = {} print(d) d[2] = 123 print(d) d = {1:2} d[3] = 3 print(len(d), d[1], d[3]) d[1] = 0 print(len(d), d[1], d[3]) print(str(d) == '{1: 0, 3: 3}' or str(d) == '{3: 3, 1: 0}') x = 1 while x < 100: d[x] = x x += 1 print(d[50]) # equality operator on dicts of different size print({} == {1:1}) # equality operator on dicts of same size but with different keys print({1:1} == {2:1}) # value not found try: {}[0] except KeyError as er: print('KeyError', er, repr(er), er.args) # unsupported unary op try: +{} except TypeError: print('TypeError') # unsupported binary op try: {} + {} except TypeError: print('TypeError')
Add test to print full KeyError exc from failed dict lookup.
tests: Add test to print full KeyError exc from failed dict lookup.
Python
mit
jmarcelino/pycom-micropython,alex-march/micropython,hiway/micropython,AriZuu/micropython,chrisdearman/micropython,kerneltask/micropython,jmarcelino/pycom-micropython,selste/micropython,tuc-osg/micropython,blazewicz/micropython,oopy/micropython,ryannathans/micropython,micropython/micropython-esp32,trezor/micropython,infinnovation/micropython,MrSurly/micropython,puuu/micropython,adafruit/micropython,torwag/micropython,pfalcon/micropython,micropython/micropython-esp32,AriZuu/micropython,ryannathans/micropython,pozetroninc/micropython,tuc-osg/micropython,pozetroninc/micropython,cwyark/micropython,dmazzella/micropython,pramasoul/micropython,tobbad/micropython,lowRISC/micropython,HenrikSolver/micropython,TDAbboud/micropython,pramasoul/micropython,infinnovation/micropython,puuu/micropython,blazewicz/micropython,SHA2017-badge/micropython-esp32,infinnovation/micropython,selste/micropython,AriZuu/micropython,adafruit/micropython,swegener/micropython,mhoffma/micropython,adafruit/circuitpython,mhoffma/micropython,oopy/micropython,deshipu/micropython,trezor/micropython,PappaPeppar/micropython,jmarcelino/pycom-micropython,pfalcon/micropython,tobbad/micropython,ryannathans/micropython,tobbad/micropython,bvernoux/micropython,chrisdearman/micropython,pozetroninc/micropython,blazewicz/micropython,HenrikSolver/micropython,hiway/micropython,torwag/micropython,ryannathans/micropython,AriZuu/micropython,henriknelson/micropython,henriknelson/micropython,mhoffma/micropython,dmazzella/micropython,PappaPeppar/micropython,Timmenem/micropython,mhoffma/micropython,blazewicz/micropython,infinnovation/micropython,oopy/micropython,tralamazza/micropython,dxxb/micropython,TDAbboud/micropython,puuu/micropython,chrisdearman/micropython,PappaPeppar/micropython,Timmenem/micropython,alex-march/micropython,pozetroninc/micropython,TDAbboud/micropython,Peetz0r/micropython-esp32,HenrikSolver/micropython,pramasoul/micropython,TDAbboud/micropython,HenrikSolver/micropython,AriZuu/micropython,oopy/micropython,alex-march/micropython,pramasoul/micropython,tobbad/micropython,alex-robbins/micropython,kerneltask/micropython,pfalcon/micropython,henriknelson/micropython,pfalcon/micropython,adafruit/circuitpython,torwag/micropython,Timmenem/micropython,cwyark/micropython,tuc-osg/micropython,tuc-osg/micropython,MrSurly/micropython,toolmacher/micropython,SHA2017-badge/micropython-esp32,henriknelson/micropython,adafruit/circuitpython,pozetroninc/micropython,micropython/micropython-esp32,alex-robbins/micropython,alex-robbins/micropython,adafruit/micropython,SHA2017-badge/micropython-esp32,dmazzella/micropython,Peetz0r/micropython-esp32,puuu/micropython,swegener/micropython,dxxb/micropython,tuc-osg/micropython,adafruit/micropython,lowRISC/micropython,MrSurly/micropython-esp32,micropython/micropython-esp32,MrSurly/micropython-esp32,hosaka/micropython,bvernoux/micropython,selste/micropython,PappaPeppar/micropython,matthewelse/micropython,matthewelse/micropython,trezor/micropython,MrSurly/micropython-esp32,hiway/micropython,SHA2017-badge/micropython-esp32,MrSurly/micropython,adafruit/circuitpython,hiway/micropython,blazewicz/micropython,kerneltask/micropython,henriknelson/micropython,Peetz0r/micropython-esp32,selste/micropython,kerneltask/micropython,MrSurly/micropython,micropython/micropython-esp32,alex-march/micropython,pfalcon/micropython,matthewelse/micropython,alex-robbins/micropython,toolmacher/micropython,puuu/micropython,toolmacher/micropython,tralamazza/micropython,torwag/micropython,hosaka/micropython,hosaka/micropython,alex-march/micropython,trezor/micropython,Timmenem/micropython,hosaka/micropython,ryannathans/micropython,swegener/micropython,jmarcelino/pycom-micropython,mhoffma/micropython,Peetz0r/micropython-esp32,dxxb/micropython,Peetz0r/micropython-esp32,swegener/micropython,toolmacher/micropython,torwag/micropython,deshipu/micropython,deshipu/micropython,adafruit/circuitpython,dxxb/micropython,lowRISC/micropython,cwyark/micropython,Timmenem/micropython,matthewelse/micropython,MrSurly/micropython-esp32,tralamazza/micropython,oopy/micropython,MrSurly/micropython,chrisdearman/micropython,dxxb/micropython,tralamazza/micropython,bvernoux/micropython,hiway/micropython,deshipu/micropython,matthewelse/micropython,toolmacher/micropython,hosaka/micropython,HenrikSolver/micropython,TDAbboud/micropython,tobbad/micropython,swegener/micropython,adafruit/circuitpython,infinnovation/micropython,cwyark/micropython,bvernoux/micropython,adafruit/micropython,trezor/micropython,MrSurly/micropython-esp32,dmazzella/micropython,lowRISC/micropython,kerneltask/micropython,SHA2017-badge/micropython-esp32,lowRISC/micropython,deshipu/micropython,chrisdearman/micropython,matthewelse/micropython,cwyark/micropython,selste/micropython,alex-robbins/micropython,PappaPeppar/micropython,jmarcelino/pycom-micropython,pramasoul/micropython,bvernoux/micropython
--- +++ @@ -26,8 +26,8 @@ # value not found try: {}[0] -except KeyError: - print('KeyError') +except KeyError as er: + print('KeyError', er, repr(er), er.args) # unsupported unary op try:
ad346d73a27c021de131ec871dc19da2e17854ee
armstrong/dev/virtualdjango/base.py
armstrong/dev/virtualdjango/base.py
import os, sys DEFAULT_SETTINGS = { 'DATABASE_ENGINE': 'sqlite3', 'DATABASES': { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'mydatabase' } }, } class VirtualDjango(object): def __init__(self, caller=sys.modules['__main__'], default_settings=DEFAULT_SETTINGS): self.caller = caller self.default_settings = default_settings def configure_settings(self, customizations, reset=True): # Django expects a `DATABASE_ENGINE` value custom_settings = self.default_settings custom_settings.update(customizations) settings = self.settings if reset: settings._wrapped = None settings.configure(**custom_settings) @property def settings(self): from django.conf import settings return settings @property def call_command(self): from django.core.management import call_command return call_command def run(self, my_settings): if hasattr(self.caller, 'setUp'): self.caller.setUp() self.configure_settings(my_settings) return self.call_command
import django import os, sys DEFAULT_SETTINGS = { 'DATABASE_ENGINE': 'sqlite3', 'DATABASES': { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'mydatabase' } }, } class VirtualDjango(object): def __init__(self, caller=sys.modules['__main__'], default_settings=DEFAULT_SETTINGS): self.caller = caller self.default_settings = default_settings def configure_settings(self, customizations, reset=True): # Django expects a `DATABASE_ENGINE` value custom_settings = self.default_settings custom_settings.update(customizations) settings = self.settings if reset: self.reset_settings(settings) settings.configure(**custom_settings) def reset_settings(self, settings): if django.VERSION[:2] == (1, 3): settings._wrapped = None return # This is the way to reset settings going forward from django.utils.functional import empty settings._wrapped = empty @property def settings(self): from django.conf import settings return settings @property def call_command(self): from django.core.management import call_command return call_command def run(self, my_settings): if hasattr(self.caller, 'setUp'): self.caller.setUp() self.configure_settings(my_settings) return self.call_command
Adjust the settings reset to work with Django 1.4
Adjust the settings reset to work with Django 1.4 Django changed the `settings._wrapped` value from `None` to the special `empty` object. This change maintains backwards compatibility for 1.3.X, while using the new method for all other versions of Django.
Python
apache-2.0
armstrong/armstrong.dev
--- +++ @@ -1,3 +1,4 @@ +import django import os, sys DEFAULT_SETTINGS = { @@ -25,8 +26,17 @@ settings = self.settings if reset: + self.reset_settings(settings) + settings.configure(**custom_settings) + + def reset_settings(self, settings): + if django.VERSION[:2] == (1, 3): settings._wrapped = None - settings.configure(**custom_settings) + return + + # This is the way to reset settings going forward + from django.utils.functional import empty + settings._wrapped = empty @property def settings(self):
1e4bdeeb1156c4ff65d261ba261f594be57bc30f
tests/test_appinfo.py
tests/test_appinfo.py
import io import os import pickle import pytest from steamfiles import appinfo test_file_name = os.path.join(os.path.dirname(__file__), 'test_data/appinfo.vdf') @pytest.yield_fixture def vdf_data(): with open(test_file_name, 'rb') as f: yield f.read() @pytest.mark.usefixtures('vdf_data') def test_load_dump(vdf_data): with open(test_file_name, 'rb') as in_file: out_file = io.BytesIO() obj = appinfo.load(in_file) appinfo.dump(obj, out_file) # Rewind to the beginning out_file.seek(0) assert out_file.read() == vdf_data def test_loads_wrong_type(): with pytest.raises(TypeError): appinfo.loads('JustTestData')
import io import os import pickle import pytest from steamfiles import appinfo test_file_name = os.path.join(os.path.dirname(__file__), 'test_data/appinfo.vdf') @pytest.yield_fixture def vdf_data(): with open(test_file_name, 'rb') as f: yield f.read() @pytest.mark.usefixtures('vdf_data') def test_loads_dumps(vdf_data): assert appinfo.dumps(appinfo.loads(vdf_data)) == vdf_data @pytest.mark.usefixtures('vdf_data') def test_load_dump(vdf_data): with open(test_file_name, 'rb') as in_file: out_file = io.BytesIO() obj = appinfo.load(in_file) appinfo.dump(obj, out_file) # Rewind to the beginning out_file.seek(0) assert out_file.read() == vdf_data def test_loads_wrong_type(): with pytest.raises(TypeError): appinfo.loads('JustTestData')
Add loads-dumps test for Appinfo
Add loads-dumps test for Appinfo
Python
mit
leovp/steamfiles
--- +++ @@ -11,6 +11,11 @@ def vdf_data(): with open(test_file_name, 'rb') as f: yield f.read() + + +@pytest.mark.usefixtures('vdf_data') +def test_loads_dumps(vdf_data): + assert appinfo.dumps(appinfo.loads(vdf_data)) == vdf_data @pytest.mark.usefixtures('vdf_data')
331f776eef9acd0509c7534040ef225869305d7f
tests/test_cookies.py
tests/test_cookies.py
# -*- coding: utf-8 -*- def test_cookies_fixture(testdir): """Make sure that pytest accepts the `cookies` fixture.""" # create a temporary pytest test module testdir.makepyfile(""" def test_valid_fixture(cookies): assert hasattr(cookies, 'bake') assert callable(cookies.bake) assert cookies.error is None assert cookies.project is None """) # run pytest with the following cmd args result = testdir.runpytest('-v') # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ '*::test_valid_fixture PASSED', ]) # make sure that that we get a '0' exit code for the testsuite assert result.ret == 0 def test_help_message(testdir): result = testdir.runpytest( '--help', ) # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ 'cookies:', ])
# -*- coding: utf-8 -*- def test_cookies_fixture(testdir): """Make sure that pytest accepts the `cookies` fixture.""" # create a temporary pytest test module testdir.makepyfile(""" def test_valid_fixture(cookies): assert hasattr(cookies, 'bake') assert callable(cookies.bake) assert cookies.exception is None assert cookies.exit_code == 0 assert cookies.project is None """) # run pytest with the following cmd args result = testdir.runpytest('-v') # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ '*::test_valid_fixture PASSED', ]) # make sure that that we get a '0' exit code for the testsuite assert result.ret == 0 def test_help_message(testdir): result = testdir.runpytest( '--help', ) # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ 'cookies:', ])
Update test for cookies fixture
Update test for cookies fixture
Python
mit
hackebrot/pytest-cookies
--- +++ @@ -9,7 +9,8 @@ def test_valid_fixture(cookies): assert hasattr(cookies, 'bake') assert callable(cookies.bake) - assert cookies.error is None + assert cookies.exception is None + assert cookies.exit_code == 0 assert cookies.project is None """)
e8bb81a8be7c76c2e1839d8315bd29f381fea4ae
enable/__init__.py
enable/__init__.py
# Copyright (c) 2007-2013 by Enthought, Inc. # All rights reserved. """ A multi-platform object drawing library. Part of the Enable project of the Enthought Tool Suite. """ __version__ = '4.3.0' __requires__ = [ 'traitsui', 'PIL', ]
# Copyright (c) 2007-2013 by Enthought, Inc. # All rights reserved. """ A multi-platform object drawing library. Part of the Enable project of the Enthought Tool Suite. """ __version__ = '4.3.0' __requires__ = [ 'traitsui', 'PIL', 'casuarius', ]
Add casuarius to the list of required packages.
Add casuarius to the list of required packages.
Python
bsd-3-clause
tommy-u/enable,tommy-u/enable,tommy-u/enable,tommy-u/enable
--- +++ @@ -8,4 +8,5 @@ __requires__ = [ 'traitsui', 'PIL', + 'casuarius', ]
072497f0032bae261e78e27daf6c8e6caeea68c1
ANN.py
ANN.py
import tensorflow as tf class ANN: def __init__(self): self.inputNodes = 7 self.hiddenNodes = 64 self.hiddenNodes2 = 32 # weight3 self.outputNodes = 1 self.x = tf.placeholder("float", shape=[None, self.inputNodes], name="sensor-input") self.W1 = tf.placeholder("float", shape=[self.inputNodes, self.hiddenNodes]) self.W2 = tf.placeholder("float", shape=[self.hiddenNodes, self.hiddenNodes2]) # weight3 self.W3 = tf.placeholder("float", shape=[self.hiddenNodes2, self.outputNodes]) # weight3 self.y = tf.tanh( tf.matmul(tf.tanh(tf.matmul(tf.tanh(tf.matmul(self.x, self.W1)), self.W2)), self.W3)) # weight3 self.session = tf.Session() init = tf.initialize_all_variables() self.session.run(init) def propagate_forward(self, car, sensor_data): return self.session.run(self.y, feed_dict={self.x: sensor_data, self.W1: car.W1, self.W2: car.W2, self.W3: car.W3}) # weight3
import tensorflow as tf class ANN: def __init__(self): self.inputNodes = 7 self.hiddenNodes = 64 self.hiddenNodes2 = 64 # weight3 self.outputNodes = 1 self.x = tf.placeholder("float", shape=[None, self.inputNodes], name="sensor-input") self.W1 = tf.placeholder("float", shape=[self.inputNodes, self.hiddenNodes]) self.W2 = tf.placeholder("float", shape=[self.hiddenNodes, self.hiddenNodes2]) # weight3 self.W3 = tf.placeholder("float", shape=[self.hiddenNodes2, self.outputNodes]) # weight3 self.y = tf.tanh( tf.matmul(tf.tanh(tf.matmul(tf.tanh(tf.matmul(self.x, self.W1)), self.W2)), self.W3)) # weight3 self.session = tf.Session() init = tf.initialize_all_variables() self.session.run(init) def propagate_forward(self, car, sensor_data): return self.session.run(self.y, feed_dict={self.x: sensor_data, self.W1: car.W1, self.W2: car.W2, self.W3: car.W3}) # weight3
Change 2nd hidden layer size
Change 2nd hidden layer size
Python
mit
DiamondOperators/selfdriving-car
--- +++ @@ -5,7 +5,7 @@ def __init__(self): self.inputNodes = 7 self.hiddenNodes = 64 - self.hiddenNodes2 = 32 # weight3 + self.hiddenNodes2 = 64 # weight3 self.outputNodes = 1 self.x = tf.placeholder("float", shape=[None, self.inputNodes], name="sensor-input")
f9ba5e64f73c3fa3fed62655c846fb4435d627cc
node/multi_var.py
node/multi_var.py
from nodes import Node class MultiVar(Node): char = "'" args = 0 results = None contents = -1 def __init__(self, node_1: Node.NodeSingle, node_2: Node.NodeSingle): self.node_1 = node_1 self.node_2 = node_2 self.args = max([node_1.args, node_2.args]) def prepare(self, stack): if len(stack) == 0: self.add_arg(stack) @Node.is_func def apply(self, *stack): self.node_2.prepare(stack) rtn = self.node_2(stack[:self.node_2.args]) self.node_1.prepare(stack) rtn.extend(self.node_1(stack[:self.node_1.args])) return rtn
from nodes import Node class MultiVar(Node): char = "'" args = 0 results = None contents = -1 def __init__(self, node_1: Node.NodeSingle, node_2: Node.NodeSingle): self.node_1 = node_1 self.node_2 = node_2 def prepare(self, stack): self.node_1.prepare(stack) self.node_2.prepare(stack) self.args = max([self.node_1.args,self.node_2.args]) @Node.is_func def apply(self, *stack): rtn = self.node_2(stack[:self.node_2.args]) rtn.extend(self.node_1(stack[:self.node_1.args])) return rtn
Fix multivar for nodes with variable lenght stacks
Fix multivar for nodes with variable lenght stacks
Python
mit
muddyfish/PYKE,muddyfish/PYKE
--- +++ @@ -10,16 +10,14 @@ def __init__(self, node_1: Node.NodeSingle, node_2: Node.NodeSingle): self.node_1 = node_1 self.node_2 = node_2 - self.args = max([node_1.args, node_2.args]) def prepare(self, stack): - if len(stack) == 0: - self.add_arg(stack) + self.node_1.prepare(stack) + self.node_2.prepare(stack) + self.args = max([self.node_1.args,self.node_2.args]) @Node.is_func def apply(self, *stack): - self.node_2.prepare(stack) rtn = self.node_2(stack[:self.node_2.args]) - self.node_1.prepare(stack) rtn.extend(self.node_1(stack[:self.node_1.args])) return rtn
253a375af742699ec4a13e5d8b33e669691ed872
mysite/urls.py
mysite/urls.py
from __future__ import print_function from cms.sitemaps import CMSSitemap from django.conf.urls import * # NOQA from django.conf.urls.i18n import i18n_patterns from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.contrib import admin from django.conf import settings admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^lab_members/', include('lab_members.urls', namespace='lab_members')), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
from __future__ import print_function from cms.sitemaps import CMSSitemap from django.conf.urls import * # NOQA from django.conf.urls.i18n import i18n_patterns from django.contrib.staticfiles.urls import staticfiles_urlpatterns from django.contrib import admin from django.conf import settings admin.autodiscover() urlpatterns = i18n_patterns('', url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), url(r'^', include('cms.urls')), ) # This is only needed when using runserver. if settings.DEBUG: urlpatterns = patterns('', url(r'^media/(?P<path>.*)$', 'django.views.static.serve', # NOQA {'document_root': settings.MEDIA_ROOT, 'show_indexes': True}), ) + staticfiles_urlpatterns() + urlpatterns # NOQA
Remove lab_members/ from project URLs
Remove lab_members/ from project URLs
Python
bsd-3-clause
mfcovington/djangocms-lab-members,mfcovington/djangocms-lab-members
--- +++ @@ -12,7 +12,6 @@ url(r'^admin/', include(admin.site.urls)), # NOQA url(r'^sitemap\.xml$', 'django.contrib.sitemaps.views.sitemap', {'sitemaps': {'cmspages': CMSSitemap}}), - url(r'^lab_members/', include('lab_members.urls', namespace='lab_members')), url(r'^', include('cms.urls')), )
e178cfd7b6abba804d8bb856a0c35a6f87209771
opentreemap/treemap/tests/ui/ui_test_urls.py
opentreemap/treemap/tests/ui/ui_test_urls.py
from django.conf.urls import patterns, include, url from django.http import HttpResponse from opentreemap import urls testing_id = 1 def full_utf8_grid(request): """ Creates a big utf8 grid where every entry is 'turned on' to point to whatever plot id is the currently assigned value of testing_id this is useful for mocking a tiler utf8 grid response so that ui tests can click the map and simulate clicking a rendered tree tile. """ global testing_id quoted_space_line = '"' + (' ' * 64) + '"' quoted_space_line_with_comma = quoted_space_line + ',' full_space_utf8_grid = ('{"grid":[' + (quoted_space_line_with_comma * 63) + quoted_space_line + '],"keys":["1"],"data":{"1":{"the_plot_id":%s}}}' % testing_id) response = HttpResponse(full_space_utf8_grid) response['Content-Type'] = 'application/json' return response urlpatterns = patterns( '', url(r'^tile/.*', full_utf8_grid), url(r'', include(urls)) )
from django.conf.urls import patterns, include, url from django.http import HttpResponse from opentreemap import urls testing_id = 1 def full_utf8_grid(request): """ Creates a big utf8 grid where every entry is 'turned on' to point to whatever plot id is the currently assigned value of testing_id this is useful for mocking a tiler utf8 grid response so that ui tests can click the map and simulate clicking a rendered tree tile. """ global testing_id quoted_space_line = '"' + (' ' * 64) + '"' quoted_space_line_with_comma = quoted_space_line + ',' full_space_utf8_grid = ('{"grid":[' + (quoted_space_line_with_comma * 63) + quoted_space_line + '],"keys":["1"],"data":{"1":{"id":%s}}}' % testing_id) response = HttpResponse(full_space_utf8_grid) response['Content-Type'] = 'application/json' return response urlpatterns = patterns( '', url(r'^tile/.*', full_utf8_grid), url(r'', include(urls)) )
Use "id" instead of "plot_id"
Use "id" instead of "plot_id" For the fake utf grid
Python
agpl-3.0
clever-crow-consulting/otm-core,recklessromeo/otm-core,RickMohr/otm-core,clever-crow-consulting/otm-core,RickMohr/otm-core,recklessromeo/otm-core,recklessromeo/otm-core,recklessromeo/otm-core,maurizi/otm-core,RickMohr/otm-core,RickMohr/otm-core,maurizi/otm-core,maurizi/otm-core,clever-crow-consulting/otm-core,maurizi/otm-core,clever-crow-consulting/otm-core
--- +++ @@ -25,7 +25,7 @@ full_space_utf8_grid = ('{"grid":[' + (quoted_space_line_with_comma * 63) + quoted_space_line + - '],"keys":["1"],"data":{"1":{"the_plot_id":%s}}}' + '],"keys":["1"],"data":{"1":{"id":%s}}}' % testing_id) response = HttpResponse(full_space_utf8_grid)
1501510e37722a2b68871d44c431ac8c44060e1e
logster/scanner.py
logster/scanner.py
import os import pymongo from .db import connect_to_db files_to_check = [ '/home/irvind/test.log', '/home/irvind/test2.log' ] def run_scanner(): _, db = connect_to_db(async=False) while True: try: pass except KeyboardInterrupt: break
import os import datetime import time import hashlib from datetime import datetime import pymongo from .db import connect_to_db def file_mtime(path): return datetime.fromtimestamp(os.stat(path).st_mtime) def md5_hex(data): # todo pass def run_scanner(): _, db = connect_to_db(async=False) try: while True: _scanner_iter(db) except KeyboardInterrupt: pass def _scanner_iter(db): def check_log(log): mtime, path, log_id = ( log.get('last_mtime'), log.get('path'), log.get('id'), ) if mtime is not None and mtime >= file_mtime(path): # File not changed return log_entries = list(db.entries.find({'log': log_id}).sort('order')) db_lines = [ent['content'] for ent in log_entries] with open(path, 'r') as f: file_lines = f.readlines() if len(file_lines) <= len(db_lines): # todo: special case return old_lines = file_lines[:len(db_lines)] new_lines = file_lines[len(db_lines):] if md5_hex(old_lines) != md5_hex(db_lines): # todo: special case return last_order = log_entries[len(db_lines)-1]['order'] new_docs = [] i = last_order + 1 for line in new_lines: new_docs.append({ 'log': log_id, 'content': line, 'order': i }) i += 1 db.entries.insert(new_docs) _notify_websockets(new_docs) for log in db.logs.find(): check_log(log) time.sleep(1) def _notify_websockets(docs): # todo pass
Monitor new records and add them to the database
[wip] Monitor new records and add them to the database
Python
mit
irvind/logster,irvind/logster,irvind/logster
--- +++ @@ -1,20 +1,88 @@ import os +import datetime +import time +import hashlib + +from datetime import datetime + import pymongo from .db import connect_to_db -files_to_check = [ - '/home/irvind/test.log', - '/home/irvind/test2.log' -] +def file_mtime(path): + return datetime.fromtimestamp(os.stat(path).st_mtime) + + +def md5_hex(data): + # todo + pass def run_scanner(): _, db = connect_to_db(async=False) + + try: + while True: + _scanner_iter(db) - while True: - try: - pass - except KeyboardInterrupt: - break + except KeyboardInterrupt: + pass + + +def _scanner_iter(db): + + def check_log(log): + mtime, path, log_id = ( + log.get('last_mtime'), + log.get('path'), + log.get('id'), + ) + + if mtime is not None and mtime >= file_mtime(path): + # File not changed + return + + log_entries = list(db.entries.find({'log': log_id}).sort('order')) + db_lines = [ent['content'] for ent in log_entries] + + with open(path, 'r') as f: + file_lines = f.readlines() + + if len(file_lines) <= len(db_lines): + # todo: special case + return + + old_lines = file_lines[:len(db_lines)] + new_lines = file_lines[len(db_lines):] + + if md5_hex(old_lines) != md5_hex(db_lines): + # todo: special case + return + + last_order = log_entries[len(db_lines)-1]['order'] + + new_docs = [] + i = last_order + 1 + for line in new_lines: + new_docs.append({ + 'log': log_id, + 'content': line, + 'order': i + }) + + i += 1 + + db.entries.insert(new_docs) + + _notify_websockets(new_docs) + + for log in db.logs.find(): + check_log(log) + + time.sleep(1) + + +def _notify_websockets(docs): + # todo + pass
5e0dde9e6bb6dc8f24c4eb47e63c5dd68ede809c
pontoon/administration/tasks.py
pontoon/administration/tasks.py
import commonware from celery.task import task from pontoon.base.models import Project from pontoon.administration.views import _update_from_repository log = commonware.log.getLogger('pontoon') @task() def update_projects_from_repository(): for project in Project.objects.all(): try: repository_type = project.repository_type repository_url = project.repository repository_path_master = project.repository_path _update_from_repository( project, repository_type, repository_url, repository_path_master) except Exception as e: log.debug('UpdateFromRepositoryTaskError: %s' % unicode(e))
import commonware import os from celery.task import task from django.conf import settings from pontoon.base.models import Project from pontoon.administration.views import _update_from_repository log = commonware.log.getLogger('pontoon') @task() def update_projects_from_repository(): for project in Project.objects.all(): try: repository_type = project.repository_type repository_url = project.repository repository_path_master = os.path.join(settings.MEDIA_ROOT, repository_type, project.name) _update_from_repository( project, repository_type, repository_url, repository_path_master) except Exception as e: log.debug('UpdateFromRepositoryTaskError: %s' % unicode(e))
Use master path as the argument, which is crucial for single-locale repositories
Use master path as the argument, which is crucial for single-locale repositories
Python
bsd-3-clause
yfdyh000/pontoon,sudheesh001/pontoon,mozilla/pontoon,m8ttyB/pontoon,mathjazz/pontoon,participedia/pontoon,sudheesh001/pontoon,Jobava/mirror-pontoon,yfdyh000/pontoon,vivekanand1101/pontoon,jotes/pontoon,m8ttyB/pontoon,yfdyh000/pontoon,mozilla/pontoon,Jobava/mirror-pontoon,participedia/pontoon,Jobava/mirror-pontoon,sudheesh001/pontoon,jotes/pontoon,mathjazz/pontoon,vivekanand1101/pontoon,mastizada/pontoon,mathjazz/pontoon,mathjazz/pontoon,m8ttyB/pontoon,jotes/pontoon,Osmose/pontoon,mastizada/pontoon,m8ttyB/pontoon,mastizada/pontoon,mozilla/pontoon,sudheesh001/pontoon,mathjazz/pontoon,mastizada/pontoon,mozilla/pontoon,vivekanand1101/pontoon,Osmose/pontoon,participedia/pontoon,yfdyh000/pontoon,Jobava/mirror-pontoon,jotes/pontoon,Osmose/pontoon,participedia/pontoon,mozilla/pontoon,Osmose/pontoon,vivekanand1101/pontoon
--- +++ @@ -1,7 +1,9 @@ import commonware +import os from celery.task import task +from django.conf import settings from pontoon.base.models import Project from pontoon.administration.views import _update_from_repository @@ -15,7 +17,9 @@ try: repository_type = project.repository_type repository_url = project.repository - repository_path_master = project.repository_path + repository_path_master = os.path.join(settings.MEDIA_ROOT, + repository_type, project.name) + _update_from_repository( project, repository_type, repository_url, repository_path_master)
e089a75174c003b4d22c4a2f8d65f3adb3d84def
feedback/models.py
feedback/models.py
from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext as _ class Feedback(models.Model): site = models.ForeignKey(Site, verbose_name=_('site')) url = models.CharField(max_length=255, verbose_name=_('url')) urlhash = models.TextField(verbose_name=_('urlhash'), default="", null=True, blank=True) useragent = models.TextField(verbose_name=_('useragent'), default="", null=True, blank=True) subject = models.CharField(max_length=255, blank=True, null=True, verbose_name=_('subject')) email = models.EmailField(blank=True, null=True, verbose_name=_('email')) text = models.TextField(verbose_name=_('text')) created = models.DateTimeField(auto_now_add=True, null=True) def __unicode__(self): return u'{url}: {subject}'.format(url=self.url, subject=self.subject)
from django.db import models from django.contrib.sites.models import Site from django.utils.translation import ugettext_lazy as _ class Feedback(models.Model): site = models.ForeignKey(Site, verbose_name=_('site')) url = models.CharField(max_length=255, verbose_name=_('url')) urlhash = models.TextField(verbose_name=_('urlhash'), default="", null=True, blank=True) useragent = models.TextField(verbose_name=_('useragent'), default="", null=True, blank=True) subject = models.CharField(max_length=255, blank=True, null=True, verbose_name=_('subject')) email = models.EmailField(blank=True, null=True, verbose_name=_('email')) text = models.TextField(verbose_name=_('text')) created = models.DateTimeField(auto_now_add=True, null=True) def __unicode__(self): return u'{url}: {subject}'.format(url=self.url, subject=self.subject)
Use ugettext_lazy instead of ugettext to support Django 1.7 migrations
Use ugettext_lazy instead of ugettext to support Django 1.7 migrations
Python
mit
PetrDlouhy/django-feedback,jaredly/django-feedback,jaredly/django-feedback,PetrDlouhy/django-feedback,PetrDlouhy/django-feedback,jaredly/django-feedback
--- +++ @@ -1,6 +1,6 @@ from django.db import models from django.contrib.sites.models import Site -from django.utils.translation import ugettext as _ +from django.utils.translation import ugettext_lazy as _ class Feedback(models.Model):
fcc9a8ed53385f4cf9ca7f1bb73fdc7407b88b7f
pryvate/blueprints/simple/simple.py
pryvate/blueprints/simple/simple.py
"""Simple blueprint.""" import os from flask import Blueprint, current_app, render_template blueprint = Blueprint('simple', __name__, url_prefix='/simple', template_folder='templates') @blueprint.route('', methods=['GET']) def get_simple(): """List all packages.""" packages = os.listdir(current_app.config['BASEDIR']) return render_template('simple.html', packages=packages) @blueprint.route('/<package>', methods=['GET']) def get_package(package): """List versions of a package.""" package_path = os.path.join(current_app.config['BASEDIR'], package.lower()) files = os.listdir(package_path) packages = [] for filename in files: if filename.endswith('md5'): with open(os.path.join(package_path, filename), 'r') as md5_digest: item = { 'name': package, 'version': filename.replace('.md5', ''), 'digest': md5_digest.read() } packages.append(item) return render_template('simple_package.html', packages=packages, letter=package[:1])
"""Simple blueprint.""" import os from flask import Blueprint, current_app, render_template blueprint = Blueprint('simple', __name__, url_prefix='/simple', template_folder='templates') @blueprint.route('', methods=['GET']) def get_simple(): """List all packages.""" packages = os.listdir(current_app.config['BASEDIR']) return render_template('simple.html', packages=packages) @blueprint.route('/<package>', methods=['GET']) def get_package(package): """List versions of a package.""" package_path = os.path.join(current_app.config['BASEDIR'], package.lower()) files = os.listdir(package_path) packages = [] for filename in files: if filename.endswith('md5'): with open(os.path.join(package_path, filename), 'r') as md5_digest: item = { 'name': package, 'version': filename.replace('.md5', ''), 'digest': md5_digest.read() } packages.append(item) return render_template('simple_package.html', packages=packages, letter=package[:1].lower())
Use only lowercase letters in the source link as well
Use only lowercase letters in the source link as well
Python
mit
Dinoshauer/pryvate,Dinoshauer/pryvate
--- +++ @@ -32,4 +32,4 @@ } packages.append(item) return render_template('simple_package.html', packages=packages, - letter=package[:1]) + letter=package[:1].lower())
ee349a7e26579ec6dffc3153fb8d55fad3281c5b
app.py
app.py
#!/usr/bin/env python from flask import Flask, abort, jsonify, request from icalendar import Calendar from urllib import urlopen app = Flask(__name__) @app.route('/') def index(): return u'Please use like <code>%swww.myserver.com/path/to/file.ics' % request.host_url @app.route('/<path:url>') def convert_from_url(url): url = 'http://%s' % url uh = urlopen(url) if uh.getcode() != 200: abort(404) ics = uh.read() uh.close() cal = Calendar.from_string(ics) data = {} data[cal.name] = dict(cal.items()) for component in cal.subcomponents: if not data[cal.name].has_key(component.name): data[cal.name][component.name] = [] comp_obj = {} for item in component.items(): comp_obj[item[0]] = unicode(item[1]) data[cal.name][component.name].append(comp_obj) resp = jsonify(data) if 'callback' in request.args: resp.data = "%s(%s);" % (request.args['callback'], resp.data) return resp if __name__ == '__main__': app.run(debug=True)
#!/usr/bin/env python from flask import Flask, abort, jsonify, request from icalendar import Calendar from urllib import urlopen app = Flask(__name__) @app.route('/') def index(): return u'Please use like <code>http://<script>document.write(location.host);</script><noscript>ical2json.ep.io</noscript>/www.myserver.com/path/to/file.ics</code>' @app.route('/<path:url>') def convert_from_url(url): url = 'http://%s' % url uh = urlopen(url) if uh.getcode() != 200: abort(404) ics = uh.read() uh.close() cal = Calendar.from_string(ics) data = {} data[cal.name] = dict(cal.items()) for component in cal.subcomponents: if not data[cal.name].has_key(component.name): data[cal.name][component.name] = [] comp_obj = {} for item in component.items(): comp_obj[item[0]] = unicode(item[1]) data[cal.name][component.name].append(comp_obj) resp = jsonify(data) if 'callback' in request.args: resp.data = "%s(%s);" % (request.args['callback'], resp.data) return resp if __name__ == '__main__': app.run(debug=True)
Fix index page example URL.
Fix index page example URL.
Python
mit
philippbosch/ical2json
--- +++ @@ -8,7 +8,7 @@ @app.route('/') def index(): - return u'Please use like <code>%swww.myserver.com/path/to/file.ics' % request.host_url + return u'Please use like <code>http://<script>document.write(location.host);</script><noscript>ical2json.ep.io</noscript>/www.myserver.com/path/to/file.ics</code>' @app.route('/<path:url>') def convert_from_url(url):
5c04957ca44fc43eae034fe389d39b879ec000ae
slideshow/event.py
slideshow/event.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import types _subscribers = {} class func: def __init__(self, inst, method): self._inst = inst self._method = method def __call__(self, *args, **kwargs): # bind method to class instance types.MethodType(self._method, self._inst)(*args, **kwargs) def listener(cls): for name, method in cls.__dict__.iteritems(): if hasattr(method, "subscribe"): # methods are still not bound, so store both the instance and the # method for late binding subscribe(method.subscribe, func(cls, method)) return cls def callback(event): def decorate(func): # func is not yet a class member, so just mark what event it want func.subscribe = event return func return decorate def trigger(event, *args, **kwargs): if event not in _subscribers: return for func in _subscribers[event]: func(*args, **kwargs) def subscribe(event, callback): if event not in _subscribers: _subscribers[event] = [] _subscribers[event].append(callback)
#!/usr/bin/env python # -*- coding: utf-8 -*- import types import cherrypy _subscribers = {} class func: def __init__(self, inst, method, cls): self._inst = inst self._method = method self._cls = cls def __call__(self, *args, **kwargs): # bind method to class instance types.MethodType(self._method, self._inst)(*args, **kwargs) def listener(cls): for name, method in cls.__dict__.iteritems(): if hasattr(method, "subscribe"): # methods are still not bound, so store both the instance and the # method for late binding subscribe(method.subscribe, func(cls, method, None)) return cls def callback(event): def decorate(func): # func is not yet a class member, so just mark what event it want func.subscribe = event return func return decorate def trigger(event, *args, **kwargs): cherrypy.engine.log('%s triggered' % event) if event not in _subscribers: return for func in _subscribers[event]: func(*args, **kwargs) def subscribe(event, callback): if not isinstance(event, basestring): raise ValueError, 'event name must be string (got %s)' % event.__class__.__name__ if event not in _subscribers: _subscribers[event] = [] _subscribers[event].append(callback)
Add logging and some minor error handling.
Add logging and some minor error handling.
Python
agpl-3.0
ext/slideshow-frontend,ext/slideshow,ext/slideshow-frontend,ext/slideshow,ext/slideshow,ext/slideshow,ext/slideshow-frontend,ext/slideshow,ext/slideshow
--- +++ @@ -2,25 +2,27 @@ # -*- coding: utf-8 -*- import types +import cherrypy _subscribers = {} class func: - def __init__(self, inst, method): + def __init__(self, inst, method, cls): self._inst = inst self._method = method - + self._cls = cls + def __call__(self, *args, **kwargs): # bind method to class instance types.MethodType(self._method, self._inst)(*args, **kwargs) def listener(cls): - for name, method in cls.__dict__.iteritems(): + for name, method in cls.__dict__.iteritems(): if hasattr(method, "subscribe"): # methods are still not bound, so store both the instance and the # method for late binding - subscribe(method.subscribe, func(cls, method)) - return cls + subscribe(method.subscribe, func(cls, method, None)) + return cls def callback(event): def decorate(func): @@ -30,14 +32,19 @@ return decorate def trigger(event, *args, **kwargs): + cherrypy.engine.log('%s triggered' % event) + if event not in _subscribers: return - + for func in _subscribers[event]: func(*args, **kwargs) def subscribe(event, callback): + if not isinstance(event, basestring): + raise ValueError, 'event name must be string (got %s)' % event.__class__.__name__ + if event not in _subscribers: _subscribers[event] = [] - + _subscribers[event].append(callback)
ea8f2c3b036ff62ea2fa7faea6eefc86ac5471c7
redis_sessions_fork/settings.py
redis_sessions_fork/settings.py
import os from django.conf import settings SESSION_REDIS_HOST = getattr( settings, 'SESSION_REDIS_HOST', '127.0.0.1' ) SESSION_REDIS_PORT = getattr( settings, 'SESSION_REDIS_PORT', 6379 ) SESSION_REDIS_DB = getattr( settings, 'SESSION_REDIS_DB', 0 ) SESSION_REDIS_PREFIX = getattr( settings, 'SESSION_REDIS_PREFIX', 'django_sessions' ) SESSION_REDIS_PASSWORD = getattr( settings, 'SESSION_REDIS_PASSWORD', None ) SESSION_REDIS_UNIX_DOMAIN_SOCKET_PATH = getattr( settings, 'SESSION_REDIS_UNIX_DOMAIN_SOCKET_PATH', None ) SESSION_REDIS_URL = getattr( settings, 'SESSION_REDIS_URL', None ) SESSION_REDIS_CONNECTION_POOL = getattr( settings, 'SESSION_REDIS_CONNECTION_POOL', None ) if SESSION_REDIS_URL is None: # redis clouds ENV variables SESSION_REDIS_ENV_URLS = getattr( settings, 'SESSION_REDIS_ENV_URLS', ( 'REDISTOGO_URL', 'OPENREDIS_URL', 'REDISGREEN_URL', 'MYREDIS_URL', 'REDISCLOUD_URL' ) ) for url in SESSION_REDIS_ENV_URLS: redis_env_url = os.environ.get(url) if redis_env_url: SESSION_REDIS_URL = redis_env_url break
import os from django.conf import settings SESSION_REDIS_HOST = getattr( settings, 'SESSION_REDIS_HOST', '127.0.0.1' ) SESSION_REDIS_PORT = getattr( settings, 'SESSION_REDIS_PORT', 6379 ) SESSION_REDIS_DB = getattr( settings, 'SESSION_REDIS_DB', 0 ) SESSION_REDIS_PREFIX = getattr( settings, 'SESSION_REDIS_PREFIX', 'django_sessions' ) SESSION_REDIS_PASSWORD = getattr( settings, 'SESSION_REDIS_PASSWORD', None ) SESSION_REDIS_UNIX_DOMAIN_SOCKET_PATH = getattr( settings, 'SESSION_REDIS_UNIX_DOMAIN_SOCKET_PATH', None ) SESSION_REDIS_URL = getattr( settings, 'SESSION_REDIS_URL', None ) SESSION_REDIS_CONNECTION_POOL = getattr( settings, 'SESSION_REDIS_CONNECTION_POOL', None ) if SESSION_REDIS_URL is None: # redis clouds ENV variables SESSION_REDIS_ENV_URLS = getattr( settings, 'SESSION_REDIS_ENV_URLS', ( 'REDISCLOUD_URL' 'REDISTOGO_URL', 'OPENREDIS_URL', 'REDISGREEN_URL', 'MYREDIS_URL', ) ) for url in SESSION_REDIS_ENV_URLS: redis_env_url = os.environ.get(url) if redis_env_url: SESSION_REDIS_URL = redis_env_url break
Reorder ENV urls redis providers.
Reorder ENV urls redis providers.
Python
bsd-3-clause
ProDG/django-redis-sessions-fork,hellysmile/django-redis-sessions-fork
--- +++ @@ -52,11 +52,11 @@ SESSION_REDIS_ENV_URLS = getattr( settings, 'SESSION_REDIS_ENV_URLS', ( + 'REDISCLOUD_URL' 'REDISTOGO_URL', 'OPENREDIS_URL', 'REDISGREEN_URL', 'MYREDIS_URL', - 'REDISCLOUD_URL' ) )
5082354efec86bb0ebb111e51c8e5a039ab7ae88
pypika/dialects.py
pypika/dialects.py
from .enums import Dialects from .queries import ( Query, QueryBuilder, ) class MySQLQuery(Query): """ Defines a query class for use with MySQL. """ @classmethod def _builder(cls): return QueryBuilder(quote_char='`', dialect=Dialects.MYSQL, wrap_union_queries=False) class VerticaQuery(Query): """ Defines a query class for use with Vertica. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.VERTICA) class OracleQuery(Query): """ Defines a query class for use with Oracle. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.ORACLE) class PostgreSQLQuery(Query): """ Defines a query class for use with PostgreSQL. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.POSTGRESQL) class RedshiftQuery(Query): """ Defines a query class for use with Amazon Redshift. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.REDSHIFT) class MSSQLQuery(Query): """ Defines a query class for use with Microsoft SQL Server. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.MSSQL) class ClickHouseQuery(Query): """ Defines a query class for use with Yandex ClickHouse. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.CLICKHOUSE)
from .enums import Dialects from .queries import ( Query, QueryBuilder, ) class MySQLQuery(Query): """ Defines a query class for use with MySQL. """ @classmethod def _builder(cls): return QueryBuilder(quote_char='`', dialect=Dialects.MYSQL, wrap_union_queries=False) class VerticaQuery(Query): """ Defines a query class for use with Vertica. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.VERTICA) class OracleQuery(Query): """ Defines a query class for use with Oracle. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.ORACLE) class PostgreSQLQuery(Query): """ Defines a query class for use with PostgreSQL. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.POSTGRESQL) class RedshiftQuery(Query): """ Defines a query class for use with Amazon Redshift. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.REDSHIFT) class MSSQLQuery(Query): """ Defines a query class for use with Microsoft SQL Server. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.MSSQL) class ClickHouseQuery(Query): """ Defines a query class for use with Yandex ClickHouse. """ @classmethod def _builder(cls): return QueryBuilder(dialect=Dialects.CLICKHOUSE, wrap_union_queries=False)
Disable union wrap for clickhouse
Disable union wrap for clickhouse
Python
apache-2.0
kayak/pypika
--- +++ @@ -72,4 +72,4 @@ @classmethod def _builder(cls): - return QueryBuilder(dialect=Dialects.CLICKHOUSE) + return QueryBuilder(dialect=Dialects.CLICKHOUSE, wrap_union_queries=False)
b3a9a4a1e451815f15dc35c9b6ec9f7b67387260
scipy/misc/tests/test_common.py
scipy/misc/tests/test_common.py
from __future__ import division, print_function, absolute_import import pytest from numpy.testing import assert_equal, assert_allclose from scipy._lib._numpy_compat import suppress_warnings from scipy.misc import pade, logsumexp, face, ascent, electrocardiogram from scipy.special import logsumexp as sc_logsumexp def test_logsumexp(): # make sure logsumexp can be imported from either scipy.misc or # scipy.special with suppress_warnings() as sup: sup.filter(DeprecationWarning, "`logsumexp` is deprecated") assert_allclose(logsumexp([0, 1]), sc_logsumexp([0, 1]), atol=1e-16) def test_pade(): # make sure scipy.misc.pade exists with suppress_warnings() as sup: sup.filter(DeprecationWarning, "`pade` is deprecated") pade([1, 2], 1) def test_face(): assert_equal(face().shape, (768, 1024, 3)) def test_ascent(): assert_equal(ascent().shape, (512, 512)) def test_electrocardiogram(): # Test shape and dtype of signal ecg = electrocardiogram() assert_equal(ecg.shape, (108000,)) assert ecg.dtype == float
from __future__ import division, print_function, absolute_import import pytest from numpy.testing import assert_equal, assert_allclose, assert_almost_equal from scipy._lib._numpy_compat import suppress_warnings from scipy.misc import pade, logsumexp, face, ascent, electrocardiogram from scipy.special import logsumexp as sc_logsumexp def test_logsumexp(): # make sure logsumexp can be imported from either scipy.misc or # scipy.special with suppress_warnings() as sup: sup.filter(DeprecationWarning, "`logsumexp` is deprecated") assert_allclose(logsumexp([0, 1]), sc_logsumexp([0, 1]), atol=1e-16) def test_pade(): # make sure scipy.misc.pade exists with suppress_warnings() as sup: sup.filter(DeprecationWarning, "`pade` is deprecated") pade([1, 2], 1) def test_face(): assert_equal(face().shape, (768, 1024, 3)) def test_ascent(): assert_equal(ascent().shape, (512, 512)) def test_electrocardiogram(): # Test shape, dtype and stats of signal ecg = electrocardiogram() assert ecg.dtype == float assert_equal(ecg.shape, (108000,)) assert_almost_equal(ecg.mean(), -0.16510875) assert_almost_equal(ecg.std(), 0.5992473991177294)
Check mean and STD of returned ECG signal
TST: Check mean and STD of returned ECG signal
Python
bsd-3-clause
Eric89GXL/scipy,jor-/scipy,andyfaff/scipy,ilayn/scipy,andyfaff/scipy,tylerjereddy/scipy,grlee77/scipy,aarchiba/scipy,arokem/scipy,lhilt/scipy,perimosocordiae/scipy,mdhaber/scipy,rgommers/scipy,matthew-brett/scipy,aeklant/scipy,endolith/scipy,rgommers/scipy,person142/scipy,ilayn/scipy,WarrenWeckesser/scipy,matthew-brett/scipy,scipy/scipy,nmayorov/scipy,aeklant/scipy,andyfaff/scipy,endolith/scipy,Stefan-Endres/scipy,aarchiba/scipy,person142/scipy,aarchiba/scipy,ilayn/scipy,Eric89GXL/scipy,person142/scipy,pizzathief/scipy,grlee77/scipy,vigna/scipy,ilayn/scipy,anntzer/scipy,scipy/scipy,pizzathief/scipy,jamestwebber/scipy,arokem/scipy,arokem/scipy,vigna/scipy,jor-/scipy,mdhaber/scipy,gertingold/scipy,ilayn/scipy,Stefan-Endres/scipy,andyfaff/scipy,gfyoung/scipy,rgommers/scipy,lhilt/scipy,andyfaff/scipy,e-q/scipy,perimosocordiae/scipy,pizzathief/scipy,tylerjereddy/scipy,endolith/scipy,WarrenWeckesser/scipy,gfyoung/scipy,person142/scipy,Eric89GXL/scipy,endolith/scipy,matthew-brett/scipy,grlee77/scipy,Stefan-Endres/scipy,aarchiba/scipy,WarrenWeckesser/scipy,Stefan-Endres/scipy,endolith/scipy,gertingold/scipy,perimosocordiae/scipy,e-q/scipy,scipy/scipy,vigna/scipy,gertingold/scipy,gfyoung/scipy,e-q/scipy,jor-/scipy,matthew-brett/scipy,perimosocordiae/scipy,aeklant/scipy,mdhaber/scipy,WarrenWeckesser/scipy,scipy/scipy,jamestwebber/scipy,anntzer/scipy,vigna/scipy,Eric89GXL/scipy,scipy/scipy,nmayorov/scipy,Stefan-Endres/scipy,gfyoung/scipy,lhilt/scipy,tylerjereddy/scipy,Eric89GXL/scipy,Eric89GXL/scipy,anntzer/scipy,nmayorov/scipy,WarrenWeckesser/scipy,WarrenWeckesser/scipy,pizzathief/scipy,gfyoung/scipy,zerothi/scipy,scipy/scipy,e-q/scipy,matthew-brett/scipy,aarchiba/scipy,pizzathief/scipy,tylerjereddy/scipy,aeklant/scipy,aeklant/scipy,jor-/scipy,perimosocordiae/scipy,vigna/scipy,e-q/scipy,arokem/scipy,anntzer/scipy,ilayn/scipy,jor-/scipy,zerothi/scipy,mdhaber/scipy,perimosocordiae/scipy,gertingold/scipy,Stefan-Endres/scipy,nmayorov/scipy,jamestwebber/scipy,anntzer/scipy,endolith/scipy,zerothi/scipy,nmayorov/scipy,person142/scipy,andyfaff/scipy,anntzer/scipy,rgommers/scipy,grlee77/scipy,rgommers/scipy,zerothi/scipy,zerothi/scipy,jamestwebber/scipy,arokem/scipy,tylerjereddy/scipy,jamestwebber/scipy,gertingold/scipy,mdhaber/scipy,zerothi/scipy,lhilt/scipy,lhilt/scipy,grlee77/scipy,mdhaber/scipy
--- +++ @@ -1,7 +1,7 @@ from __future__ import division, print_function, absolute_import import pytest -from numpy.testing import assert_equal, assert_allclose +from numpy.testing import assert_equal, assert_allclose, assert_almost_equal from scipy._lib._numpy_compat import suppress_warnings from scipy.misc import pade, logsumexp, face, ascent, electrocardiogram @@ -32,7 +32,9 @@ def test_electrocardiogram(): - # Test shape and dtype of signal + # Test shape, dtype and stats of signal ecg = electrocardiogram() + assert ecg.dtype == float assert_equal(ecg.shape, (108000,)) - assert ecg.dtype == float + assert_almost_equal(ecg.mean(), -0.16510875) + assert_almost_equal(ecg.std(), 0.5992473991177294)
1ee41f5439f80af139e612591d48cdac5ecfda39
hiapi/hi.py
hiapi/hi.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import argparse from flask import Flask app = Flask(__name__) @app.route('/') def hello(): return 'Hi!\n' def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('-b', '--bind-address', dest='bind', default='127.0.0.1') parser.add_argument('-p', '--port', dest='port', default=4000, type=int) return parser.parse_args() def main(): opts = parse_args() app.run(host=opts.bind, port=opts.port) # Support for uWSGI def application(env, start_response): start_response('200 OK', [('Content-Type', 'text/plain')]) return [b'Hi!\n'] if __name__ == "__main__": main()
#!/usr/bin/env python # -*- coding: utf-8 -*- import argparse import flask RESPONSE_CODE = 200 app = flask.Flask(__name__) @app.route('/') def hello(): global RESPONSE_CODE if RESPONSE_CODE == 200: return 'Hi!\n' else: flask.abort(RESPONSE_CODE) def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('-b', '--bind-address', dest='bind', default='127.0.0.1') parser.add_argument('-p', '--port', dest='port', default=4000, type=int) parser.add_argument('-c', '--response_code', dest='code', default=200, type=int) return parser.parse_args() def main(): global RESPONSE_CODE opts = parse_args() RESPONSE_CODE = opts.code app.run(host=opts.bind, port=opts.port) if __name__ == "__main__": main()
Remove uwsgi support, add support for simple alternative responses
Remove uwsgi support, add support for simple alternative responses
Python
apache-2.0
GradysGhost/pyhiapi
--- +++ @@ -2,29 +2,33 @@ # -*- coding: utf-8 -*- import argparse -from flask import Flask +import flask -app = Flask(__name__) +RESPONSE_CODE = 200 + +app = flask.Flask(__name__) @app.route('/') def hello(): - return 'Hi!\n' + global RESPONSE_CODE + if RESPONSE_CODE == 200: + return 'Hi!\n' + else: + flask.abort(RESPONSE_CODE) def parse_args(): parser = argparse.ArgumentParser() parser.add_argument('-b', '--bind-address', dest='bind', default='127.0.0.1') parser.add_argument('-p', '--port', dest='port', default=4000, type=int) + parser.add_argument('-c', '--response_code', dest='code', default=200, type=int) return parser.parse_args() def main(): + global RESPONSE_CODE opts = parse_args() + RESPONSE_CODE = opts.code app.run(host=opts.bind, port=opts.port) - -# Support for uWSGI -def application(env, start_response): - start_response('200 OK', [('Content-Type', 'text/plain')]) - return [b'Hi!\n'] if __name__ == "__main__": main()
39b4141ebd939bde649a9dfeb4f4cc832d0f4c39
solutions/problem_9/solution.py
solutions/problem_9/solution.py
import sys import os sys.path.append(os.path.abspath(os.path.dirname('./utils/python'))) from python.decorators import timeit @timeit def solution(): n = 1000 for b in xrange(1, n): for c in xrange(1, n): a = n - b - c if (a**2) + (b**2) == (c**2): return a * b * c solution()
import sys import os sys.path.append(os.path.abspath(os.path.dirname('./utils/python'))) from python.utils import timeit @timeit def solution(): n = 1000 for b in xrange(1, n): for c in xrange(1, n): a = n - b - c if (a**2) + (b**2) == (c**2): return a * b * c solution()
Fix python import for problem 9
Fix python import for problem 9
Python
mit
mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler,mdsrosa/project_euler
--- +++ @@ -1,7 +1,7 @@ import sys import os sys.path.append(os.path.abspath(os.path.dirname('./utils/python'))) -from python.decorators import timeit +from python.utils import timeit @timeit
403bd1cdea0a8d1fae25710a48dc3148fc21ddd9
bell.py
bell.py
#!/usr/bin/env python from time import sleep import subprocess import httplib, urllib import RPi.GPIO as GPIO import config import logger GPIO.setmode(GPIO.BCM) GPIO.setup(config.bell_pin, GPIO.IN) LOW_PRIORITY = -1 MEDIUM_PRIORITY = 0 HIGH_PRIORITY = 1 log = logger.get(__name__) def notifyPhones(message, priority=MEDIUM_PRIORITY): log.debug('Sending pushover message "' + message + '" with priority ' + str(priority)) conn = httplib.HTTPSConnection("api.pushover.net:443") conn.request("POST", "/1/messages.json", urllib.urlencode({ "token": config.application_token, "user": config.user_token, "title": config.message_title, "message": message, "url": config.message_url, "priority": priority, }), { "Content-type": "application/x-www-form-urlencoded" }) response = conn.getresponse() log.debug('Got response: ' + str(response.status) + ' ' + response.reason + ': ' + response.read()) conn.close() notifyPhones('Listener started', LOW_PRIORITY) log.info('Doorbell listener Started') while True: if (GPIO.input(23) == False): subprocess.Popen(["ogg123","-q","dingdong.ogg"]) notifyPhones(config.message_text) log.info('Doorbell pressed') sleep(3);
#!/usr/bin/env python import logging import RPi.GPIO as GPIO from application import logsetup, button, pushover GPIO.setmode(GPIO.BCM) GPIO.setup(config.bell_pin, GPIO.IN) log = logging.getLogger(__name__) log.info('Doorbell listener Started') pushover.send('Listener started', pushover.LOW_PRIORITY) while True: if (GPIO.input(23) == False): button.pressed()
Convert Pi version to use application modules
Convert Pi version to use application modules
Python
mit
viv/pibell
--- +++ @@ -1,50 +1,17 @@ #!/usr/bin/env python -from time import sleep -import subprocess -import httplib, urllib +import logging import RPi.GPIO as GPIO -import config -import logger +from application import logsetup, button, pushover GPIO.setmode(GPIO.BCM) GPIO.setup(config.bell_pin, GPIO.IN) -LOW_PRIORITY = -1 -MEDIUM_PRIORITY = 0 -HIGH_PRIORITY = 1 +log = logging.getLogger(__name__) -log = logger.get(__name__) - -def notifyPhones(message, priority=MEDIUM_PRIORITY): - log.debug('Sending pushover message "' - + message - + '" with priority ' - + str(priority)) - conn = httplib.HTTPSConnection("api.pushover.net:443") - conn.request("POST", "/1/messages.json", - urllib.urlencode({ - "token": config.application_token, - "user": config.user_token, - "title": config.message_title, - "message": message, - "url": config.message_url, - "priority": priority, - }), { "Content-type": "application/x-www-form-urlencoded" }) - - response = conn.getresponse() - log.debug('Got response: ' - + str(response.status) - + ' ' + response.reason - + ': ' + response.read()) - conn.close() - -notifyPhones('Listener started', LOW_PRIORITY) log.info('Doorbell listener Started') +pushover.send('Listener started', pushover.LOW_PRIORITY) while True: if (GPIO.input(23) == False): - subprocess.Popen(["ogg123","-q","dingdong.ogg"]) - notifyPhones(config.message_text) - log.info('Doorbell pressed') - sleep(3); + button.pressed()
dde6e451a9e434b980d1ebac84626ec7515485c5
instruments/bbn.py
instruments/bbn.py
from .instrument import Instrument, VisaInterface from types import MethodType class Attenuator(Instrument): NUM_CHANNELS = 3 """BBN 3 Channel Instrument""" def __init__(self, name, resource_name): super(Attenuator, self).__init__(name, resource_name, interface_type="VISA") self.interface._resource.baud_rate = 115200 self.interface._resource.read_termination = u"\r\n" self.interface._resource.write_termination = u"\n" #Clear "unknown command" from connect #TODO: where the heck does this come from # self.interface.read() # self.interface.read() #Override query to look for ``end`` def query(self, query_string): val = self._resource.query(query_string) assert self.read() == "END" return val self.interface.query = MethodType(query, self.interface, VisaInterface) def get_attenuation(self, chan): return float(self.interface.query("GET {:d}".format(chan))) def set_attenuation(self, chan, val): self.interface.write("SET {:d} {:.1f}".format(chan, val)) assert self.interface.read() == "Setting channel {:d} to {:.2f}".format(chan, val) assert self.interface.read() == "END"
from .instrument import Instrument, VisaInterface from types import MethodType class Attenuator(Instrument): """BBN 3 Channel Instrument""" NUM_CHANNELS = 3 def __init__(self, name, resource_name): super(Attenuator, self).__init__(name, resource_name, interface_type="VISA") self.interface._resource.baud_rate = 115200 self.interface._resource.read_termination = u"\r\n" self.interface._resource.write_termination = u"\n" #Override query to look for ``end`` def query(self, query_string): val = self._resource.query(query_string) assert self.read() == "END" return val self.interface.query = MethodType(query, self.interface, VisaInterface) @classmethod def channel_check(cls, chan): """ Assert the channel requested is feasbile """ assert chan > 0 and chan <= cls.NUM_CHANNELS, "Invalid channel requested: channel ({:d}) must be between 1 and {:d}".format(chan, cls.NUM_CHANNELS) def get_attenuation(self, chan): Attenuator.channel_check(chan) return float(self.interface.query("GET {:d}".format(chan))) def set_attenuation(self, chan, val): Attenuator.channel_check(chan) self.interface.write("SET {:d} {:.1f}".format(chan, val)) assert self.interface.read() == "Setting channel {:d} to {:.2f}".format(chan, val) assert self.interface.read() == "END"
Add some channel validity checking to digital attenuator
Add some channel validity checking to digital attenuator --CAR
Python
apache-2.0
BBN-Q/Auspex,BBN-Q/Auspex,BBN-Q/Auspex,BBN-Q/Auspex
--- +++ @@ -1,20 +1,17 @@ from .instrument import Instrument, VisaInterface from types import MethodType + class Attenuator(Instrument): + """BBN 3 Channel Instrument""" + NUM_CHANNELS = 3 - """BBN 3 Channel Instrument""" def __init__(self, name, resource_name): super(Attenuator, self).__init__(name, resource_name, interface_type="VISA") self.interface._resource.baud_rate = 115200 self.interface._resource.read_termination = u"\r\n" self.interface._resource.write_termination = u"\n" - - #Clear "unknown command" from connect - #TODO: where the heck does this come from - # self.interface.read() - # self.interface.read() #Override query to look for ``end`` def query(self, query_string): @@ -24,10 +21,17 @@ self.interface.query = MethodType(query, self.interface, VisaInterface) + @classmethod + def channel_check(cls, chan): + """ Assert the channel requested is feasbile """ + assert chan > 0 and chan <= cls.NUM_CHANNELS, "Invalid channel requested: channel ({:d}) must be between 1 and {:d}".format(chan, cls.NUM_CHANNELS) + def get_attenuation(self, chan): + Attenuator.channel_check(chan) return float(self.interface.query("GET {:d}".format(chan))) def set_attenuation(self, chan, val): + Attenuator.channel_check(chan) self.interface.write("SET {:d} {:.1f}".format(chan, val)) assert self.interface.read() == "Setting channel {:d} to {:.2f}".format(chan, val) assert self.interface.read() == "END"
cbd39a43d8da9b9fa81582d7e9aaf299631022ae
pytoon/main.py
pytoon/main.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from pytoon.connection import BrickConnection app = Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///app.db' app.config['SQLALCHEMY_COMMIT_ON_TEARDOWN'] = True db = SQLAlchemy(app) print(db.get_app()) print('starting database') db.create_all() print('database created') @app.route('/') def index(): timestamps = Electricity.query.all() data = '<table><tbody>' for t in timestamps: data += '<tr><td>{}</td></tr>'.format(t.timestamp) data += '</tbody></table>' return data class PyToon(object): def __init__(self, database): host = "192.168.178.35" port = 4223 BrickConnection(host, port, database) class Electricity(db.Model): timestamp = db.Column(db.DateTime, primary_key=True) def __repr__(self): return '<Timestamp {}>'.format(self.timestamp) if __name__ == '__main__': pt = PyToon(db) app.run(debug=True, use_reloader=False)
#!/usr/bin/env python # -*- coding: utf-8 -*- import os from flask import Flask from flask.ext.sqlalchemy import SQLAlchemy from pytoon.connection import BrickConnection app = Flask(__name__) app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:////tmp/app.db' app.config['SQLALCHEMY_COMMIT_ON_TEARDOWN'] = True db = SQLAlchemy(app) print(db.get_app()) print('starting database') print('database created') @app.route('/') def index(): timestamps = Electricity.query.all() data = '<table><tbody>' for t in timestamps: data += '<tr><td>{}</td></tr>'.format(t.timestamp) data += '</tbody></table>' return data class PyToon(object): def __init__(self, database): host = "192.168.178.35" port = 4223 BrickConnection(host, port, database) class Electricity(db.Model): timestamp = db.Column(db.DateTime, primary_key=True) def __init__(self, timestamp): self.timestamp = timestamp def __repr__(self): return '<Timestamp {}>'.format(self.timestamp) db.create_all() if __name__ == '__main__': pt = PyToon(db) app.run(debug=True, use_reloader=False)
Fix problem with creating tables
Fix problem with creating tables
Python
bsd-3-clause
marcofinalist/pytoon,marcofinalist/pytoon,marcoplaisier/pytoon,marcoplaisier/pytoon
--- +++ @@ -7,12 +7,12 @@ app = Flask(__name__) -app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:///app.db' +app.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:////tmp/app.db' app.config['SQLALCHEMY_COMMIT_ON_TEARDOWN'] = True db = SQLAlchemy(app) print(db.get_app()) print('starting database') -db.create_all() + print('database created') @@ -36,8 +36,13 @@ class Electricity(db.Model): timestamp = db.Column(db.DateTime, primary_key=True) + def __init__(self, timestamp): + self.timestamp = timestamp + def __repr__(self): return '<Timestamp {}>'.format(self.timestamp) + +db.create_all() if __name__ == '__main__': pt = PyToon(db)
3a21d7d174dd8a9cbf76cbf149666337a8352c61
h5py/_hl/compat.py
h5py/_hl/compat.py
""" Compatibility module for high-level h5py """ import sys from os import fspath, fsencode, fsdecode WINDOWS_ENCODING = "mbcs" def filename_encode(filename): """ Encode filename for use in the HDF5 library. Due to how HDF5 handles filenames on different systems, this should be called on any filenames passed to the HDF5 library. See the documentation on filenames in h5py for more information. """ filename = fspath(filename) if sys.platform == "win32": if isinstance(filename, str): return filename.encode(WINDOWS_ENCODING, "strict") return filename return fsencode(filename) def filename_decode(filename): """ Decode filename used by HDF5 library. Due to how HDF5 handles filenames on different systems, this should be called on any filenames passed from the HDF5 library. See the documentation on filenames in h5py for more information. """ if sys.platform == "win32": if isinstance(filename, bytes): return filename.decode(WINDOWS_ENCODING, "strict") elif isinstance(filename, str): return filename else: raise TypeError("expect bytes or str, not %s" % type(filename).__name__) return fsdecode(filename)
""" Compatibility module for high-level h5py """ import sys from os import fspath, fsencode, fsdecode from ..version import hdf5_built_version_tuple WINDOWS_ENCODING = "utf-8" if hdf5_built_version_tuple >= (1, 10, 6) else "mbcs" def filename_encode(filename): """ Encode filename for use in the HDF5 library. Due to how HDF5 handles filenames on different systems, this should be called on any filenames passed to the HDF5 library. See the documentation on filenames in h5py for more information. """ filename = fspath(filename) if sys.platform == "win32": if isinstance(filename, str): return filename.encode(WINDOWS_ENCODING, "strict") return filename return fsencode(filename) def filename_decode(filename): """ Decode filename used by HDF5 library. Due to how HDF5 handles filenames on different systems, this should be called on any filenames passed from the HDF5 library. See the documentation on filenames in h5py for more information. """ if sys.platform == "win32": if isinstance(filename, bytes): return filename.decode(WINDOWS_ENCODING, "strict") elif isinstance(filename, str): return filename else: raise TypeError("expect bytes or str, not %s" % type(filename).__name__) return fsdecode(filename)
Use UTF-8 on windows where HDF5 >= 1.10.6
Use UTF-8 on windows where HDF5 >= 1.10.6
Python
bsd-3-clause
h5py/h5py,h5py/h5py,h5py/h5py
--- +++ @@ -3,8 +3,9 @@ """ import sys from os import fspath, fsencode, fsdecode +from ..version import hdf5_built_version_tuple -WINDOWS_ENCODING = "mbcs" +WINDOWS_ENCODING = "utf-8" if hdf5_built_version_tuple >= (1, 10, 6) else "mbcs" def filename_encode(filename):
35fcaad0474df3352ccdf0545fc34cf2c431761c
tweet_s3_images.py
tweet_s3_images.py
import exifread import os class TweetS3Images(object): def __init__(self, twitter, s3_client): self._twitter = twitter self._s3_client = s3_client self._file = None def send_image(self, bucket, image_name, cleanup=False): temp_file = '/tmp/{}'.format(image_name) self._s3_client.download_file(bucket, image_name, temp_file) self._file = open(temp_file, 'rb') status = 'New image {} brought to you by lambda-tweet'.format(image_name) tags = exifread.process_file(self._file) self._twitter.update_with_media(filename=image_name, status=status, file=self._file) if cleanup: self.cleanup(temp_file) def get_file(self): return self._file def cleanup(self, file): os.remove(file)
import exifread import os class TweetS3Images(object): def __init__(self, twitter, s3_client): self._twitter = twitter self._s3_client = s3_client self._file = None def send_image(self, bucket, image_name, cleanup=False): temp_file = '/tmp/{}'.format(image_name) self._s3_client.download_file(bucket, image_name, temp_file) self._file = open(temp_file, 'rb') tags = exifread.process_file(self._file) status = self.get_image_description(tags, image_name) self._twitter.update_with_media(filename=image_name, status=status, file=self._file) if cleanup: self.cleanup(temp_file) def get_file(self): return self._file @staticmethod def cleanup(file_to_remove): os.remove(file_to_remove) @staticmethod def get_image_description(tags, image_name): if 'Image ImageDescription' in tags: status = tags['Image ImageDescription'].values else: status = 'New image {} brought to you by lambda-tweet'.format(image_name) return status
Update class to get image description if available and use it as the update message.
Update class to get image description if available and use it as the update message.
Python
mit
onema/lambda-tweet
--- +++ @@ -12,8 +12,8 @@ temp_file = '/tmp/{}'.format(image_name) self._s3_client.download_file(bucket, image_name, temp_file) self._file = open(temp_file, 'rb') - status = 'New image {} brought to you by lambda-tweet'.format(image_name) tags = exifread.process_file(self._file) + status = self.get_image_description(tags, image_name) self._twitter.update_with_media(filename=image_name, status=status, file=self._file) @@ -23,5 +23,15 @@ def get_file(self): return self._file - def cleanup(self, file): - os.remove(file) + @staticmethod + def cleanup(file_to_remove): + os.remove(file_to_remove) + + @staticmethod + def get_image_description(tags, image_name): + if 'Image ImageDescription' in tags: + status = tags['Image ImageDescription'].values + else: + status = 'New image {} brought to you by lambda-tweet'.format(image_name) + + return status
b096d91564366392c4003b26bafd1e6c3fff47d3
trayapp.py
trayapp.py
# Github Tray App import rumps import config import contribs username = config.get_username() class GithubTrayApp(rumps.App): @rumps.timer(60*5) def timer(self, sender): count = contribs.get_contribs(username) self.title = str(count) @rumps.clicked('Update') def onoff(self, sender): count = contribs.get_contribs(username) self.title = str(count) @rumps.clicked('Change Frequency') def onoff(self, sender): rumps.alert('jk! not ready yet!') @rumps.clicked('Change Username') def prefs(self, _): rumps.alert('jk! not ready yet!') if __name__ == "__main__": count = contribs.get_contribs(username) GithubTrayApp('Github', icon='github.png', title=str(count)).run()
# Github Tray App import rumps import config import contribs class GithubTrayApp(rumps.App): def __init__(self): super(GithubTrayApp, self).__init__('Github') self.count = rumps.MenuItem('commits') self.username = config.get_username() self.icon = 'github.png' self.menu = [ self.count, 'Update Now', 'Change Frequency', 'Change Username' ] self.update() def update(self): num = contribs.get_contribs(self.username) self.count.title = str(num) + ' commits' self.title = str(num) @rumps.timer(60*5) def timer(self, _): print('Running timer') self.update() @rumps.clicked('Update Now') def update_now(self, _): self.update() @rumps.clicked('Change Frequency') def change_frequency(_): rumps.alert('jk! not ready yet!') @rumps.clicked('Change Username') def change_username(_): rumps.alert('jk! not ready yet!') if __name__ == '__main__': GithubTrayApp().run()
Restructure app and display commit count as a disabled menu item
Restructure app and display commit count as a disabled menu item
Python
mit
chrisfosterelli/commitwatch
--- +++ @@ -5,28 +5,42 @@ import config import contribs -username = config.get_username() - class GithubTrayApp(rumps.App): + def __init__(self): + super(GithubTrayApp, self).__init__('Github') + self.count = rumps.MenuItem('commits') + self.username = config.get_username() + self.icon = 'github.png' + self.menu = [ + self.count, + 'Update Now', + 'Change Frequency', + 'Change Username' + ] + self.update() + + def update(self): + num = contribs.get_contribs(self.username) + self.count.title = str(num) + ' commits' + self.title = str(num) + @rumps.timer(60*5) - def timer(self, sender): - count = contribs.get_contribs(username) - self.title = str(count) + def timer(self, _): + print('Running timer') + self.update() - @rumps.clicked('Update') - def onoff(self, sender): - count = contribs.get_contribs(username) - self.title = str(count) + @rumps.clicked('Update Now') + def update_now(self, _): + self.update() @rumps.clicked('Change Frequency') - def onoff(self, sender): + def change_frequency(_): rumps.alert('jk! not ready yet!') @rumps.clicked('Change Username') - def prefs(self, _): + def change_username(_): rumps.alert('jk! not ready yet!') -if __name__ == "__main__": - count = contribs.get_contribs(username) - GithubTrayApp('Github', icon='github.png', title=str(count)).run() +if __name__ == '__main__': + GithubTrayApp().run()
b8d3e62bae3559b24a2a135c921ccc9879fab339
src/py65/memory.py
src/py65/memory.py
class ObservableMemory: def __init__(self, subject=None): if subject is None: subject = 0x10000 * [0x00] self._subject = subject self._read_subscribers = {} self._write_subscribers = {} def __setitem__(self, address, value): callbacks = self._write_subscribers.setdefault(address, []) for callback in callbacks: result = callback(address, value) if result is not None: value = result self._subject[address] = value def __getitem__(self, address): callbacks = self._read_subscribers.setdefault(address, []) final_result = None for callback in callbacks: result = callback(address) if result is not None: final_result = result if final_result is None: return self._subject[address] else: return final_result def __getattr__(self, attribute): return getattr(self._subject, attribute) def subscribe_to_write(self, address_range, callback): for address in address_range: callbacks = self._write_subscribers.setdefault(address, []) if callback not in callbacks: callbacks.append(callback) def subscribe_to_read(self, address_range, callback): for address in address_range: callbacks = self._read_subscribers.setdefault(address, []) if callback not in callbacks: callbacks.append(callback) def write(self, start_address, bytes): self._subject[start_address:start_address+len(bytes)] = bytes
class ObservableMemory: def __init__(self, subject=None): if subject is None: subject = 0x10000 * [0x00] self._subject = subject self._read_subscribers = {} self._write_subscribers = {} def __setitem__(self, address, value): callbacks = self._write_subscribers.get(address, []) for callback in callbacks: result = callback(address, value) if result is not None: value = result self._subject[address] = value def __getitem__(self, address): callbacks = self._read_subscribers.get(address, []) final_result = None for callback in callbacks: result = callback(address) if result is not None: final_result = result if final_result is None: return self._subject[address] else: return final_result def __getattr__(self, attribute): return getattr(self._subject, attribute) def subscribe_to_write(self, address_range, callback): for address in address_range: callbacks = self._write_subscribers.setdefault(address, []) if callback not in callbacks: callbacks.append(callback) def subscribe_to_read(self, address_range, callback): for address in address_range: callbacks = self._read_subscribers.setdefault(address, []) if callback not in callbacks: callbacks.append(callback) def write(self, start_address, bytes): self._subject[start_address:start_address+len(bytes)] = bytes
Use get() instead of setdefault().
Use get() instead of setdefault().
Python
bsd-3-clause
mkeller0815/py65,mnaberez/py65
--- +++ @@ -9,7 +9,7 @@ self._write_subscribers = {} def __setitem__(self, address, value): - callbacks = self._write_subscribers.setdefault(address, []) + callbacks = self._write_subscribers.get(address, []) for callback in callbacks: result = callback(address, value) @@ -19,7 +19,7 @@ self._subject[address] = value def __getitem__(self, address): - callbacks = self._read_subscribers.setdefault(address, []) + callbacks = self._read_subscribers.get(address, []) final_result = None for callback in callbacks:
f6e42f0f2d931533a9cdcab749feceeb5fa98982
mopidy/backends/spotify/__init__.py
mopidy/backends/spotify/__init__.py
"""A backend for playing music from Spotify `Spotify <http://www.spotify.com/>`_ is a music streaming service. The backend uses the official `libspotify <http://developer.spotify.com/en/libspotify/overview/>`_ library and the `pyspotify <http://github.com/mopidy/pyspotify/>`_ Python bindings for libspotify. This backend handles URIs starting with ``spotify:``. See :ref:`music-from-spotify` for further instructions on using this backend. .. note:: This product uses SPOTIFY(R) CORE but is not endorsed, certified or otherwise approved in any way by Spotify. Spotify is the registered trade mark of the Spotify Group. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Spotify+backend **Dependencies:** - libspotify >= 11, < 12 (libspotify11 package from apt.mopidy.com) - pyspotify >= 1.7, < 1.8 (python-spotify package from apt.mopidy.com) **Settings:** - :attr:`mopidy.settings.SPOTIFY_CACHE_PATH` - :attr:`mopidy.settings.SPOTIFY_USERNAME` - :attr:`mopidy.settings.SPOTIFY_PASSWORD` """ # flake8: noqa from .actor import SpotifyBackend
"""A backend for playing music from Spotify `Spotify <http://www.spotify.com/>`_ is a music streaming service. The backend uses the official `libspotify <http://developer.spotify.com/en/libspotify/overview/>`_ library and the `pyspotify <http://github.com/mopidy/pyspotify/>`_ Python bindings for libspotify. This backend handles URIs starting with ``spotify:``. See :ref:`music-from-spotify` for further instructions on using this backend. .. note:: This product uses SPOTIFY(R) CORE but is not endorsed, certified or otherwise approved in any way by Spotify. Spotify is the registered trade mark of the Spotify Group. **Issues:** https://github.com/mopidy/mopidy/issues?labels=Spotify+backend **Dependencies:** - libspotify >= 12, < 13 (libspotify12 package from apt.mopidy.com) - pyspotify >= 1.8, < 1.9 (python-spotify package from apt.mopidy.com) **Settings:** - :attr:`mopidy.settings.SPOTIFY_CACHE_PATH` - :attr:`mopidy.settings.SPOTIFY_USERNAME` - :attr:`mopidy.settings.SPOTIFY_PASSWORD` """ # flake8: noqa from .actor import SpotifyBackend
Update recommended libspotify and pyspotify version
Update recommended libspotify and pyspotify version
Python
apache-2.0
SuperStarPL/mopidy,bacontext/mopidy,priestd09/mopidy,mokieyue/mopidy,hkariti/mopidy,liamw9534/mopidy,adamcik/mopidy,mokieyue/mopidy,rawdlite/mopidy,dbrgn/mopidy,kingosticks/mopidy,mokieyue/mopidy,mopidy/mopidy,SuperStarPL/mopidy,glogiotatidis/mopidy,adamcik/mopidy,rawdlite/mopidy,pacificIT/mopidy,kingosticks/mopidy,hkariti/mopidy,liamw9534/mopidy,swak/mopidy,glogiotatidis/mopidy,tkem/mopidy,dbrgn/mopidy,bencevans/mopidy,ali/mopidy,adamcik/mopidy,quartz55/mopidy,ZenithDK/mopidy,pacificIT/mopidy,hkariti/mopidy,ZenithDK/mopidy,swak/mopidy,tkem/mopidy,rawdlite/mopidy,pacificIT/mopidy,woutervanwijk/mopidy,glogiotatidis/mopidy,jmarsik/mopidy,jmarsik/mopidy,abarisain/mopidy,vrs01/mopidy,jcass77/mopidy,diandiankan/mopidy,rawdlite/mopidy,diandiankan/mopidy,glogiotatidis/mopidy,ali/mopidy,jmarsik/mopidy,diandiankan/mopidy,abarisain/mopidy,jcass77/mopidy,jodal/mopidy,jcass77/mopidy,bacontext/mopidy,hkariti/mopidy,tkem/mopidy,jodal/mopidy,tkem/mopidy,priestd09/mopidy,mopidy/mopidy,SuperStarPL/mopidy,mokieyue/mopidy,bencevans/mopidy,priestd09/mopidy,pacificIT/mopidy,bacontext/mopidy,swak/mopidy,mopidy/mopidy,swak/mopidy,bacontext/mopidy,diandiankan/mopidy,jmarsik/mopidy,ZenithDK/mopidy,ZenithDK/mopidy,vrs01/mopidy,jodal/mopidy,vrs01/mopidy,dbrgn/mopidy,quartz55/mopidy,quartz55/mopidy,dbrgn/mopidy,SuperStarPL/mopidy,bencevans/mopidy,bencevans/mopidy,ali/mopidy,kingosticks/mopidy,vrs01/mopidy,quartz55/mopidy,ali/mopidy,woutervanwijk/mopidy
--- +++ @@ -20,8 +20,8 @@ **Dependencies:** -- libspotify >= 11, < 12 (libspotify11 package from apt.mopidy.com) -- pyspotify >= 1.7, < 1.8 (python-spotify package from apt.mopidy.com) +- libspotify >= 12, < 13 (libspotify12 package from apt.mopidy.com) +- pyspotify >= 1.8, < 1.9 (python-spotify package from apt.mopidy.com) **Settings:**
87c99c6839d7f74201393066651f3d69a5a2edce
nodeconductor/logging/middleware.py
nodeconductor/logging/middleware.py
from __future__ import unicode_literals import threading _locals = threading.local() def get_event_context(): return getattr(_locals, 'context', None) def set_event_context(context): _locals.context = context def reset_event_context(): if hasattr(_locals, 'context'): del _locals.context def get_ip_address(request): """ Correct IP address is expected as first element of HTTP_X_FORWARDED_FOR or REMOTE_ADDR """ if 'HTTP_X_FORWARDED_FOR' in request.META: return request.META['HTTP_X_FORWARDED_FOR'].split(',')[0].strip() else: return request.META['REMOTE_ADDR'] class CaptureEventContextMiddleware(object): def process_request(self, request): context = {'ip_address': get_ip_address(request)} user = getattr(request, 'user', None) if user and not user.is_anonymous(): context.update(user._get_log_context('user')) set_event_context(context) def process_response(self, request, response): reset_event_context() return response
from __future__ import unicode_literals import threading _locals = threading.local() def get_event_context(): return getattr(_locals, 'context', None) def set_event_context(context): _locals.context = context def reset_event_context(): if hasattr(_locals, 'context'): del _locals.context def set_current_user(user): set_event_context(user._get_log_context('user')) def get_ip_address(request): """ Correct IP address is expected as first element of HTTP_X_FORWARDED_FOR or REMOTE_ADDR """ if 'HTTP_X_FORWARDED_FOR' in request.META: return request.META['HTTP_X_FORWARDED_FOR'].split(',')[0].strip() else: return request.META['REMOTE_ADDR'] class CaptureEventContextMiddleware(object): def process_request(self, request): context = {'ip_address': get_ip_address(request)} user = getattr(request, 'user', None) if user and not user.is_anonymous(): context.update(user._get_log_context('user')) set_event_context(context) def process_response(self, request, response): reset_event_context() return response
Implement set_current_user for CapturingAuthentication (NC-529)
Implement set_current_user for CapturingAuthentication (NC-529)
Python
mit
opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor
--- +++ @@ -16,6 +16,10 @@ def reset_event_context(): if hasattr(_locals, 'context'): del _locals.context + + +def set_current_user(user): + set_event_context(user._get_log_context('user')) def get_ip_address(request):
6183f41cf6fa2125d7d5de626a7b2a994253d4ca
src/levels/main.py
src/levels/main.py
# Copyright 2013 Daniel Stokes, Mitchell Stokes # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import os from bge import logic def init(): if ".." not in sys.path: os.chdir(logic.expandPath("//")) sys.path.append("..") try: from scripts.state import StartupState, DefaultState from scripts.framework.state import StateSystem if ('dostartup' in sys.argv): logic.state_system = StateSystem(StartupState) else: print("Skipping startup") logic.state_system = StateSystem(DefaultState) except: import traceback traceback.print_exc() logic.state_system = None def run(): try: logic.state_system except AttributeError: init() if not logic.state_system: return logic.state_system.update()
# Copyright 2013 Daniel Stokes, Mitchell Stokes # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import sys import os from bge import logic def init(): if ".." not in sys.path: os.chdir(logic.expandPath("//../")) sys.path.append(".") try: from scripts.state import StartupState, DefaultState from scripts.framework.state import StateSystem if ('dostartup' in sys.argv): logic.state_system = StateSystem(StartupState) else: print("Skipping startup") logic.state_system = StateSystem(DefaultState) except: import traceback traceback.print_exc() logic.state_system = None def run(): try: logic.state_system except AttributeError: init() if not logic.state_system: return logic.state_system.update()
Use the src directory instead of the src/levels directory as the cwd.
Use the src directory instead of the src/levels directory as the cwd.
Python
apache-2.0
Moguri/ullur,Moguri/ullur
--- +++ @@ -20,8 +20,8 @@ def init(): if ".." not in sys.path: - os.chdir(logic.expandPath("//")) - sys.path.append("..") + os.chdir(logic.expandPath("//../")) + sys.path.append(".") try: from scripts.state import StartupState, DefaultState
3da6393345cca10d44c0823ef6c224a5ceaa4fcd
src/engine/SCons/Platform/darwin.py
src/engine/SCons/Platform/darwin.py
"""engine.SCons.Platform.darwin Platform-specific initialization for Mac OS X systems. There normally shouldn't be any need to import this module directly. It will usually be imported through the generic SCons.Platform.Platform() selection method. """ # # Copyright (c) 2001, 2002, 2003, 2004 Steven Knight # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be included # in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # import posix import os def generate(env): posix.generate(env) env['SHLIBSUFFIX'] = '.dylib'
"""engine.SCons.Platform.darwin Platform-specific initialization for Mac OS X systems. There normally shouldn't be any need to import this module directly. It will usually be imported through the generic SCons.Platform.Platform() selection method. """ # # __COPYRIGHT__ # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be included # in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # __revision__ = "__FILE__ __REVISION__ __DATE__ __DEVELOPER__" import posix import os def generate(env): posix.generate(env) env['SHLIBSUFFIX'] = '.dylib'
Fix __COPYRIGHT__ and __REVISION__ in new Darwin module.
Fix __COPYRIGHT__ and __REVISION__ in new Darwin module.
Python
mit
azatoth/scons,azatoth/scons,azatoth/scons,azatoth/scons,azatoth/scons
--- +++ @@ -8,7 +8,7 @@ """ # -# Copyright (c) 2001, 2002, 2003, 2004 Steven Knight +# __COPYRIGHT__ # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the @@ -30,6 +30,8 @@ # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # +__revision__ = "__FILE__ __REVISION__ __DATE__ __DEVELOPER__" + import posix import os
6480e801de5f21486d99444c25006b70329e580e
luigi/tasks/release/process_data.py
luigi/tasks/release/process_data.py
# -*- coding: utf-8 -*- """ Copyright [2009-2017] EMBL-European Bioinformatics Institute Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import luigi from tasks.ena import Ena from tasks.ensembl.ensembl import Ensembl from tasks.pdb import Pdb from tasks.rfam import RfamSequences from tasks.rfam import RfamFamilies class ProcessData(luigi.WrapperTask): # pylint: disable=R0904 """ This will generate the CSV's to import for all the databases we update each release. """ def requires(self): yield Ena() yield Ensembl() yield RfamSequences() yield RfamFamilies() yield Pdb()
# -*- coding: utf-8 -*- """ Copyright [2009-2017] EMBL-European Bioinformatics Institute Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ import luigi from tasks.ena import Ena from tasks.ensembl.ensembl import Ensembl from tasks.pdb import Pdb from tasks.rfam import RfamSequences from tasks.rfam import RfamFamilies from tasks.rgd import Rgd class ProcessData(luigi.WrapperTask): # pylint: disable=R0904 """ This will generate the CSV's to import for all the databases we update each release. """ def requires(self): yield Ena() yield Ensembl() yield RfamSequences() yield RfamFamilies() yield Pdb() yield Rgd()
Add RGD as part of the regular update pipeline
Add RGD as part of the regular update pipeline I'm not sure if this will always be part of the update, but it is for at least for this release.
Python
apache-2.0
RNAcentral/rnacentral-import-pipeline,RNAcentral/rnacentral-import-pipeline,RNAcentral/rnacentral-import-pipeline,RNAcentral/rnacentral-import-pipeline
--- +++ @@ -20,6 +20,7 @@ from tasks.pdb import Pdb from tasks.rfam import RfamSequences from tasks.rfam import RfamFamilies +from tasks.rgd import Rgd class ProcessData(luigi.WrapperTask): # pylint: disable=R0904 @@ -34,3 +35,4 @@ yield RfamSequences() yield RfamFamilies() yield Pdb() + yield Rgd()
57a1299e771b97c81249cd3075ff9e7e047356a6
wikked/__init__.py
wikked/__init__.py
from flask import Flask, abort # Create the main app. app = Flask(__name__) app.config.from_object('wikked.settings') app.config.from_envvar('WIKKED_SETTINGS', silent=True) if app.config['DEBUG']: from werkzeug import SharedDataMiddleware import os app.wsgi_app = SharedDataMiddleware(app.wsgi_app, { '/': os.path.join(os.path.dirname(__file__), 'static') }) # The main Wiki instance. from wiki import Wiki wiki = Wiki(logger=app.logger) # Import views and user loader. import wikked.views # Login extension. from flask.ext.login import LoginManager login_manager = LoginManager() login_manager.init_app(app) login_manager.user_loader(wiki.auth.getUser) login_manager.unauthorized_handler(lambda: abort(401)) # Bcrypt extension. from flaskext.bcrypt import Bcrypt app.bcrypt = Bcrypt(app)
from flask import Flask, abort # Create the main app. app = Flask(__name__) app.config.from_object('wikked.settings') app.config.from_envvar('WIKKED_SETTINGS', silent=True) if app.config['DEBUG']: from werkzeug import SharedDataMiddleware import os app.wsgi_app = SharedDataMiddleware(app.wsgi_app, { '/': os.path.join(os.path.dirname(__file__), 'static') }) # The main Wiki instance. from wiki import Wiki wiki = Wiki(root=app.config.get('WIKI_ROOT'), logger=app.logger) # Import views and user loader. import wikked.views # Login extension. from flask.ext.login import LoginManager login_manager = LoginManager() login_manager.init_app(app) login_manager.user_loader(wiki.auth.getUser) login_manager.unauthorized_handler(lambda: abort(401)) # Bcrypt extension. from flaskext.bcrypt import Bcrypt app.bcrypt = Bcrypt(app)
Use the settings' wiki root, if any.
Use the settings' wiki root, if any.
Python
apache-2.0
ludovicchabant/Wikked,ludovicchabant/Wikked,ludovicchabant/Wikked
--- +++ @@ -14,7 +14,7 @@ # The main Wiki instance. from wiki import Wiki -wiki = Wiki(logger=app.logger) +wiki = Wiki(root=app.config.get('WIKI_ROOT'), logger=app.logger) # Import views and user loader. import wikked.views
08d8f2c30c810cda75961e5bf6025f1bf348fc02
api.py
api.py
from os import environ from eve import Eve from settings import API_NAME api = Eve(API_NAME) if __name__ == '__main__': # Heroku support: bind to PORT if defined, otherwise default to 5000. if 'PORT' in environ: port = int(environ.get('PORT')) host = '0.0.0.0' else: port = 5000 host = '127.0.0.1' api.run(host=host, port=port)
import json from os import environ from eve import Eve from settings import API_NAME, URL_PREFIX api = Eve(API_NAME) def add_document(resource, document): "Add a new document to the given resource." return api.test_client().post('/' + URL_PREFIX + '/' + resource, data=json.dumps(document), content_type='application/json') if __name__ == '__main__': # Heroku support: bind to PORT if defined, otherwise default to 5000. if 'PORT' in environ: port = int(environ.get('PORT')) host = '0.0.0.0' else: port = 5000 host = '127.0.0.1' api.run(host=host, port=port)
Add utility method to add documents
Add utility method to add documents
Python
apache-2.0
gwob/Maarifa,gwob/Maarifa,gwob/Maarifa,gwob/Maarifa,gwob/Maarifa
--- +++ @@ -1,10 +1,18 @@ +import json from os import environ from eve import Eve -from settings import API_NAME +from settings import API_NAME, URL_PREFIX api = Eve(API_NAME) + + +def add_document(resource, document): + "Add a new document to the given resource." + return api.test_client().post('/' + URL_PREFIX + '/' + resource, + data=json.dumps(document), + content_type='application/json') if __name__ == '__main__': # Heroku support: bind to PORT if defined, otherwise default to 5000.
a93607eea32fd5e56de47bf6c7e0e3f098d7bcd8
runTwircBot.py
runTwircBot.py
#!/usr/bin/env python from src.TwircBot import TwircBot import sys try: bot = TwircBot(sys.argv[1]) except IndexError: bot = TwircBot() bot.print_config() bot.start()
#!/usr/bin/env python3 from src.TwircBot import TwircBot import sys try: bot = TwircBot(sys.argv[1]) except IndexError: bot = TwircBot() bot.print_config() bot.start()
Change shebang to use python3 instead of python
Change shebang to use python3 instead of python
Python
mit
johnmarcampbell/twircBot
--- +++ @@ -1,4 +1,4 @@ -#!/usr/bin/env python +#!/usr/bin/env python3 from src.TwircBot import TwircBot import sys
dd9161c772e3c345fd21f742b09a62d43f7fa069
scripts/c19.py
scripts/c19.py
from __future__ import print_function import sys from pyspark.sql import SparkSession from pyspark.sql.functions import lit, col if __name__ == "__main__": if len(sys.argv) != 3: print("Usage: c19.py <input> <output>", file=sys.stderr) exit(-1) spark = SparkSession.builder.appName('Select c19').getOrCreate() raw = spark.read.option('mergeSchema','true').load(sys.argv[1]) df = raw.filter(col('date') < '1900') opens = df.filter(col('open') == 'true')\ .select('series', 'date', lit(1).alias('inopen')).distinct() df.join(opens, ['series', 'date'], 'left_outer')\ .filter((col('open') == 'true') | col('inopen').isNull())\ .drop('inopen')\ .dropDuplicates(['id'])\ .write.save(sys.argv[2]) spark.stop()
from __future__ import print_function import sys from pyspark.sql import SparkSession from pyspark.sql.functions import lit, col, datediff if __name__ == "__main__": if len(sys.argv) != 3: print("Usage: c19.py <input> <output>", file=sys.stderr) exit(-1) spark = SparkSession.builder.appName('Select c19').getOrCreate() raw = spark.read.option('mergeSchema','true').load(sys.argv[1]) df = raw.filter(col('date') < '1900') opens = df.filter(col('open') == 'true')\ .select('series', 'date', lit(1).alias('inopen')).distinct() df.join(opens, ['series', 'date'], 'left_outer')\ .filter((col('open') == 'true') | col('inopen').isNull())\ .drop('inopen')\ .dropDuplicates(['id'])\ .withColumn('ednum', col('ed').cast('int')).na.fill(1, ['ednum']) \ .withColumn('hour', datediff(col('date'), lit('1970-01-01')) * 24 + col('ednum') - 1) \ .drop('ednum') \ .write.save(sys.argv[2]) spark.stop()
Add numeric time column (hour) that respects editions.
Add numeric time column (hour) that respects editions.
Python
apache-2.0
ViralTexts/vt-passim,ViralTexts/vt-passim,ViralTexts/vt-passim
--- +++ @@ -3,7 +3,7 @@ import sys from pyspark.sql import SparkSession -from pyspark.sql.functions import lit, col +from pyspark.sql.functions import lit, col, datediff if __name__ == "__main__": if len(sys.argv) != 3: @@ -20,6 +20,9 @@ .filter((col('open') == 'true') | col('inopen').isNull())\ .drop('inopen')\ .dropDuplicates(['id'])\ + .withColumn('ednum', col('ed').cast('int')).na.fill(1, ['ednum']) \ + .withColumn('hour', datediff(col('date'), lit('1970-01-01')) * 24 + col('ednum') - 1) \ + .drop('ednum') \ .write.save(sys.argv[2]) spark.stop()
fa67de4900be765a5ea4194b1a786cd237934a33
displacy_service_tests/test_server.py
displacy_service_tests/test_server.py
import falcon.testing import json from displacy_service.server import APP class TestAPI(falcon.testing.TestCase): def __init__(self): self.api = APP def test_deps(): test_api = TestAPI() result = test_api.simulate_post(path='/dep', body='''{"text": "This is a test.", "model": "en", "collapse_punctuation": false, "collapse_phrases": false}''') result = json.loads(result.text) words = [w['text'] for w in result['words']] assert words == ["This", "is", "a", "test", "."] def test_ents(): test_api = TestAPI() result = test_api.simulate_post(path='/ent', body='''{"text": "What a great company Google is.", "model": "en"}''') ents = json.loads(result.text) assert ents == [ {"start": 21, "end": 27, "type": "ORG", "text": "Google"}] def test_sents(): test_api = TestAPI() sentences = test_api.simulate_post( path='/sent', body='''{"text": "This a test that should split into sentences! This is the second. Is this the third?", "model": "en"}''') assert sentences == ['This a test that should split into sentences!', 'This is the second.', 'Is this the third?']
import falcon.testing import json from displacy_service.server import APP class TestAPI(falcon.testing.TestCase): def __init__(self): self.api = APP def test_deps(): test_api = TestAPI() result = test_api.simulate_post( path='/dep', body='''{"text": "This is a test.", "model": "en", "collapse_punctuation": false, "collapse_phrases": false}''' ) result = json.loads(result.text) words = [w['text'] for w in result['words']] assert words == ["This", "is", "a", "test", "."] def test_ents(): test_api = TestAPI() result = test_api.simulate_post( path='/ent', body='''{"text": "What a great company Google is.", "model": "en"}''') ents = json.loads(result.text) assert ents == [ {"start": 21, "end": 27, "type": "ORG", "text": "Google"}] def test_sents(): test_api = TestAPI() sentences = test_api.simulate_post( path='/sent', body='''{"text": "This a test that should split into sentences! This is the second. Is this the third?", "model": "en"}''' ) assert sentences == ['This a test that should split into sentences!', 'This is the second.', 'Is this the third?']
Make test file PEP8 compliant.
Make test file PEP8 compliant.
Python
mit
jgontrum/spacy-api-docker,jgontrum/spacy-api-docker,jgontrum/spacy-api-docker,jgontrum/spacy-api-docker
--- +++ @@ -11,10 +11,12 @@ def test_deps(): test_api = TestAPI() - result = test_api.simulate_post(path='/dep', - body='''{"text": "This is a test.", "model": "en", - "collapse_punctuation": false, - "collapse_phrases": false}''') + result = test_api.simulate_post( + path='/dep', + body='''{"text": "This is a test.", "model": "en", + "collapse_punctuation": false, + "collapse_phrases": false}''' + ) result = json.loads(result.text) words = [w['text'] for w in result['words']] assert words == ["This", "is", "a", "test", "."] @@ -22,8 +24,9 @@ def test_ents(): test_api = TestAPI() - result = test_api.simulate_post(path='/ent', - body='''{"text": "What a great company Google is.", + result = test_api.simulate_post( + path='/ent', + body='''{"text": "What a great company Google is.", "model": "en"}''') ents = json.loads(result.text) assert ents == [ @@ -33,5 +36,10 @@ def test_sents(): test_api = TestAPI() sentences = test_api.simulate_post( - path='/sent', body='''{"text": "This a test that should split into sentences! This is the second. Is this the third?", "model": "en"}''') - assert sentences == ['This a test that should split into sentences!', 'This is the second.', 'Is this the third?'] + path='/sent', + body='''{"text": "This a test that should split into sentences! + This is the second. Is this the third?", "model": "en"}''' + ) + + assert sentences == ['This a test that should split into sentences!', + 'This is the second.', 'Is this the third?']
5db0af8fcf83519a44ed59d14bb00bc08e2a5131
django_seo_js/middleware/useragent.py
django_seo_js/middleware/useragent.py
import re from django_seo_js import settings from django_seo_js.backends import SelectedBackend from django_seo_js.helpers import request_should_be_ignored import logging logger = logging.getLogger(__name__) class UserAgentMiddleware(SelectedBackend): def __init__(self, *args, **kwargs): super(UserAgentMiddleware, self).__init__(*args, **kwargs) regex_str = "|".join(settings.USER_AGENTS) regex_str = ".*?(%s)" % regex_str self.USER_AGENT_REGEX = re.compile(regex_str, re.IGNORECASE) def process_request(self, request): if not hasattr(request, 'ENABLED') or not request.ENABLED: return if request_should_be_ignored(request): return if "HTTP_USER_AGENT" not in request.META: return if not self.USER_AGENT_REGEX.match(request.META["HTTP_USER_AGENT"]): return url = request.build_absolute_uri() try: return self.backend.get_response_for_url(url) except Exception as e: logger.exception(e)
import re from django_seo_js import settings from django_seo_js.backends import SelectedBackend from django_seo_js.helpers import request_should_be_ignored import logging logger = logging.getLogger(__name__) class UserAgentMiddleware(SelectedBackend): def __init__(self, *args, **kwargs): super(UserAgentMiddleware, self).__init__(*args, **kwargs) regex_str = "|".join(settings.USER_AGENTS) regex_str = ".*?(%s)" % regex_str self.USER_AGENT_REGEX = re.compile(regex_str, re.IGNORECASE) def process_request(self, request): if not settings.ENABLED: return if request_should_be_ignored(request): return if "HTTP_USER_AGENT" not in request.META: return if not self.USER_AGENT_REGEX.match(request.META["HTTP_USER_AGENT"]): return url = request.build_absolute_uri() try: return self.backend.get_response_for_url(url) except Exception as e: logger.exception(e)
Change from request.ENABLED to settings.ENABLED
Change from request.ENABLED to settings.ENABLED
Python
mit
skoczen/django-seo-js
--- +++ @@ -15,7 +15,7 @@ self.USER_AGENT_REGEX = re.compile(regex_str, re.IGNORECASE) def process_request(self, request): - if not hasattr(request, 'ENABLED') or not request.ENABLED: + if not settings.ENABLED: return if request_should_be_ignored(request):
5ff2390e462219f7dfa148c6ab1bdcecd9248484
app.py
app.py
from flask import Flask, render_template, request import requests from requests_oauthlib import OAuth1 app = Flask(__name__) app.config.from_pyfile('app.config') @app.route('/') def index(): term = request.args.get('q') icons = [] if term: auth = OAuth1( app.config['OAUTH_KEY'], client_secret=app.config['OAUTH_SECRET'] ) url = "http://api.thenounproject.com/icons/{0}?limit_to_public_domain=1".format(term) response = requests.get(url, auth=auth) if response.ok: icons = response.json().get('icons', []) else: term='' return render_template('index.html', icons=icons, query=term) if __name__ == '__main__': app.run(debug=True)
from flask import Flask, render_template, request import requests from requests_oauthlib import OAuth1 app = Flask(__name__) app.jinja_env.trim_blocks = True app.jinja_env.lstrip_blocks = True app.config.from_pyfile('app.config') @app.route('/') def index(): term = request.args.get('q') icons = [] if term: auth = OAuth1( app.config['OAUTH_KEY'], client_secret=app.config['OAUTH_SECRET'] ) url = "http://api.thenounproject.com/icons/{0}?limit_to_public_domain=1".format(term) response = requests.get(url, auth=auth) if response.ok: icons = response.json().get('icons', []) else: term='' return render_template('index.html', icons=icons, query=term) if __name__ == '__main__': app.run(debug=True)
Trim whitespace on rendered templates
Trim whitespace on rendered templates
Python
unlicense
ecarreras/pdicons,ecarreras/pdicons,ecarreras/pdicons,ecarreras/pdicons
--- +++ @@ -3,6 +3,8 @@ from requests_oauthlib import OAuth1 app = Flask(__name__) +app.jinja_env.trim_blocks = True +app.jinja_env.lstrip_blocks = True app.config.from_pyfile('app.config')
7729af745b9894cfb07d99f754465934a90da40d
gigs/settings_prod.py
gigs/settings_prod.py
from gigs.settings_base import * DEBUG = False TEMPLATE_DEBUG = False ADMINS = () INSTALLED_APPS += ( 'gunicorn', ) LOGGING = { 'version': 1, 'disable_existing_loggers': True, 'formatters': { 'standard': { 'format': '%(asctime)s [%(levelname)s] %(name)s: %(message)s' }, }, 'handlers': { 'default': { 'level':'DEBUG', 'class':'logging.handlers.RotatingFileHandler', 'filename': '/var/log/tugg/tugg.log', 'maxBytes': 1024*1024*50, # 50 MB 'backupCount': 5, 'formatter':'standard', }, 'request_handler': { 'level':'DEBUG', 'class':'logging.handlers.RotatingFileHandler', 'filename': '/var/log/tugg/tugg_request.log', 'maxBytes': 1024*1024*50, # 5 MB 'backupCount': 5, 'formatter':'standard', }, }, 'loggers': { '': { 'handlers': ['default'], 'level': 'DEBUG', 'propagate': True }, 'django.request': { 'handlers': ['request_handler'], 'level': 'DEBUG', 'propagate': False }, }, }
from gigs.settings_base import * DEBUG = False TEMPLATE_DEBUG = False ADMINS = () INSTALLED_APPS += ( 'gunicorn', ) DATABASES = { 'default': { 'ENGINE': 'django.db.backends.mysql', 'NAME': 'django', 'OPTIONS':{ 'read_default_file':'/etc/tugg/my.cnf', }, }, } LOGGING = { 'version': 1, 'disable_existing_loggers': True, 'formatters': { 'standard': { 'format': '%(asctime)s [%(levelname)s] %(name)s: %(message)s' }, }, 'handlers': { 'default': { 'level':'DEBUG', 'class':'logging.handlers.RotatingFileHandler', 'filename': '/var/log/tugg/tugg.log', 'maxBytes': 1024*1024*50, # 50 MB 'backupCount': 5, 'formatter':'standard', }, 'request_handler': { 'level':'DEBUG', 'class':'logging.handlers.RotatingFileHandler', 'filename': '/var/log/tugg/tugg_request.log', 'maxBytes': 1024*1024*50, # 5 MB 'backupCount': 5, 'formatter':'standard', }, }, 'loggers': { '': { 'handlers': ['default'], 'level': 'DEBUG', 'propagate': True }, 'django.request': { 'handlers': ['request_handler'], 'level': 'DEBUG', 'propagate': False }, }, }
Add databse config to production settings file
Add databse config to production settings file
Python
bsd-3-clause
shaunokeefe/gigs,shaunokeefe/gigs
--- +++ @@ -8,6 +8,16 @@ INSTALLED_APPS += ( 'gunicorn', ) + +DATABASES = { + 'default': { + 'ENGINE': 'django.db.backends.mysql', + 'NAME': 'django', + 'OPTIONS':{ + 'read_default_file':'/etc/tugg/my.cnf', + }, + }, +} LOGGING = { 'version': 1,
9826c49225a3d8aac5ab5432e261babaa2585c1e
PRESUBMIT.py
PRESUBMIT.py
# Copyright (c) 2020, the Dart project authors. Please see the AUTHORS file # for details. All rights reserved. Use of this source code is governed by a # BSD-style license that can be found in the LICENSE file. """Presubmit script for dart_ci repository. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into git cl. """ import subprocess def _NeedsFormat(path): return subprocess.call(['dartfmt', '--set-exit-if-changed', '-n', path]) != 0 def _CheckDartFormat(input_api, output_api): files = [ git_file.AbsoluteLocalPath() for git_file in input_api.AffectedTextFiles() ] unformatted_files = [ path for path in files if path.endswith('.dart') and _NeedsFormat(path) ] if unformatted_files: escapedNewline = ' \\\n' return [ output_api.PresubmitError( 'File output does not match dartfmt.\n' 'Fix these issues with:\n' 'dartfmt -w%s%s' % (escapedNewline, escapedNewline.join(unformatted_files))) ] return [] def CommonChecks(input_api, output_api): return _CheckDartFormat(input_api, output_api) CheckChangeOnCommit = CommonChecks CheckChangeOnUpload = CommonChecks
# Copyright (c) 2020, the Dart project authors. Please see the AUTHORS file # for details. All rights reserved. Use of this source code is governed by a # BSD-style license that can be found in the LICENSE file. """Presubmit script for dart_ci repository. See http://dev.chromium.org/developers/how-tos/depottools/presubmit-scripts for more details about the presubmit API built into git cl. """ import subprocess def _NeedsFormat(path): return subprocess.call(['dart', 'format', '--set-exit-if-changed', '--output','none', path]) != 0 def _CheckDartFormat(input_api, output_api): files = [ git_file.AbsoluteLocalPath() for git_file in input_api.AffectedTextFiles() ] unformatted_files = [ path for path in files if path.endswith('.dart') and _NeedsFormat(path) ] if unformatted_files: escapedNewline = ' \\\n' return [ output_api.PresubmitError( 'File output does not match dartfmt.\n' 'Fix these issues with:\n' 'dart format%s%s' % (escapedNewline, escapedNewline.join(unformatted_files))) ] return [] def CommonChecks(input_api, output_api): return _CheckDartFormat(input_api, output_api) CheckChangeOnCommit = CommonChecks CheckChangeOnUpload = CommonChecks
Update presubmit to use 'dart format' command
Update presubmit to use 'dart format' command The presubmit checks will then work with the Flutter SDK in the path. This is the case when working on the Flutter web app current_results_ui. Change-Id: I8a8d5db4454b57bc7936197032460e877037b386 Reviewed-on: https://dart-review.googlesource.com/c/dart_ci/+/191921 Reviewed-by: Alexander Thomas <29642742b6693024c89de8232f2e2542cf7eedf7@google.com>
Python
bsd-3-clause
dart-lang/dart_ci,dart-lang/dart_ci,dart-lang/dart_ci,dart-lang/dart_ci
--- +++ @@ -11,7 +11,8 @@ def _NeedsFormat(path): - return subprocess.call(['dartfmt', '--set-exit-if-changed', '-n', path]) != 0 + return subprocess.call(['dart', 'format', '--set-exit-if-changed', + '--output','none', path]) != 0 def _CheckDartFormat(input_api, output_api): @@ -29,7 +30,7 @@ output_api.PresubmitError( 'File output does not match dartfmt.\n' 'Fix these issues with:\n' - 'dartfmt -w%s%s' % + 'dart format%s%s' % (escapedNewline, escapedNewline.join(unformatted_files))) ] return []
f64fabb83cf57e70f938f803ee0a50599f3ab83a
src/odin/fields/future.py
src/odin/fields/future.py
from __future__ import absolute_import from enum import Enum from typing import TypeVar, Optional, Any, Type # noqa from odin.exceptions import ValidationError from . import Field __all__ = ("EnumField",) ET = TypeVar("ET", Enum, Enum) class EnumField(Field): """ Field for handling Python enums. """ data_type_name = "Enum" def __init__(self, enum, **options): # type: (Type[ET], **Any) -> None # Generate choices structure from choices choices = options.pop("choices", None) options["choices"] = tuple((e, e.name) for e in choices or enum) super(EnumField, self).__init__(**options) self.enum = enum @property def choices_doc_text(self): """ Choices converted for documentation purposes. """ return tuple((v.value, n) for v, n in self.choices) def to_python(self, value): # type: (Any) -> Optional[ET] if value is None: return # Attempt to convert try: return self.enum(value) except ValueError: # If value is an empty string return None # Do this check here to support enums that define an option using # an empty string. if value == "": return raise ValidationError(self.error_messages["invalid_choice"] % value) def prepare(self, value): # type: (Optional[ET]) -> Any if (value is not None) and (value in self.enum): return value.value
from __future__ import absolute_import from enum import Enum from typing import TypeVar, Optional, Any, Type # noqa from odin.exceptions import ValidationError from . import Field __all__ = ("EnumField",) ET = TypeVar("ET", Enum, Enum) class EnumField(Field): """ Field for handling Python enums. """ data_type_name = "Enum" def __init__(self, enum, **options): # type: (Type[ET], **Any) -> None # Generate choices structure from choices choices = options.pop("choices", None) options["choices"] = tuple((e, e.name) for e in choices or enum) super(EnumField, self).__init__(**options) self.enum = enum @property def choices_doc_text(self): """ Choices converted for documentation purposes. """ return tuple((v.value, n) for v, n in self.choices) def to_python(self, value): # type: (Any) -> Optional[ET] if value is None: return # Attempt to convert try: return self.enum(value) except ValueError: # If value is an empty string return None # Do this check here to support enums that define an option using # an empty string. if value == "": return raise ValidationError(self.error_messages["invalid_choice"] % value) def prepare(self, value): # type: (Optional[ET]) -> Any if (value is not None) and isinstance(value, self.enum): return value.value
Update enum value to an is-instance check
Update enum value to an is-instance check
Python
bsd-3-clause
python-odin/odin
--- +++ @@ -54,5 +54,5 @@ def prepare(self, value): # type: (Optional[ET]) -> Any - if (value is not None) and (value in self.enum): + if (value is not None) and isinstance(value, self.enum): return value.value
c3cf9adc2428c4058817daa1aeefca300363e21f
glanerbeard/server.py
glanerbeard/server.py
import requests import json import logging from glanerbeard import show log = logging.getLogger(__name__) class Server: def __init__(self, name, url, apikey): self.name = name self.url = url self.apikey = apikey def requestJson(self, path): url = '{url}/api/{apikey}{path}'.format(url=self.url,apikey=self.apikey,path=path) response = requests.get(url) response_str = response.content.decode('utf-8') result = json.loads(response_str) return result def getShows(self): shows = show.fromJson(self.requestJson('/?cmd=shows')) for s in shows: s.addServer(self) return shows def __repr__(self): return 'Server {name} at {url}'.format(name=self.name,url=self.url) def fromConfig(serverdict, apikeydict): result = [] for name,url in serverdict.items(): result.append(Server(name, url, apikeydict[name])) return result
import requests import logging from glanerbeard import show log = logging.getLogger(__name__) class Server: def __init__(self, name, url, apikey): self.name = name self.url = url self.apikey = apikey def requestJson(self, path): url = '{url}/api/{apikey}{path}'.format(url=self.url,apikey=self.apikey,path=path) return requests.get(url, verify=False).json() def getShows(self): shows = show.fromJson(self.requestJson('/?cmd=shows')) for s in shows: s.addServer(self) return shows def __repr__(self): return 'Server {name} at {url}'.format(name=self.name,url=self.url) def fromConfig(serverdict, apikeydict): result = [] for name,url in serverdict.items(): result.append(Server(name, url, apikeydict[name])) return result
Use requests built-in .json(), pass verify=False.
Use requests built-in .json(), pass verify=False.
Python
apache-2.0
daenney/glanerbeard
--- +++ @@ -1,5 +1,4 @@ import requests -import json import logging from glanerbeard import show log = logging.getLogger(__name__) @@ -12,10 +11,7 @@ def requestJson(self, path): url = '{url}/api/{apikey}{path}'.format(url=self.url,apikey=self.apikey,path=path) - response = requests.get(url) - response_str = response.content.decode('utf-8') - result = json.loads(response_str) - return result + return requests.get(url, verify=False).json() def getShows(self): shows = show.fromJson(self.requestJson('/?cmd=shows'))
321924fff843896fc67d3a4594d635546cf90bec
mycli/packages/expanded.py
mycli/packages/expanded.py
from .tabulate import _text_type def pad(field, total, char=u" "): return field + (char * (total - len(field))) def get_separator(num, header_len, data_len): sep = u"***************************[ %d. row ]***************************\n" % (num + 1) return sep def expanded_table(rows, headers): header_len = max([len(x) for x in headers]) max_row_len = 0 results = [] padded_headers = [pad(x, header_len) + u" |" for x in headers] header_len += 2 for row in rows: row_len = max([len(_text_type(x)) for x in row]) row_result = [] if row_len > max_row_len: max_row_len = row_len for header, value in zip(padded_headers, row): row_result.append(u"%s %s" % (header, value)) results.append('\n'.join(row_result)) output = [] for i, result in enumerate(results): output.append(get_separator(i, header_len, max_row_len)) output.append(result) output.append('\n') return ''.join(output)
from .tabulate import _text_type def pad(field, total, char=u" "): return field + (char * (total - len(field))) def get_separator(num, header_len, data_len): sep = u"***************************[ %d. row ]***************************\n" % (num + 1) return sep def expanded_table(rows, headers): header_len = max([len(x) for x in headers]) max_row_len = 0 results = [] padded_headers = [pad(x, header_len) + u" |" for x in headers] header_len += 2 for row in rows: row_len = max([len(_text_type(x)) for x in row]) row_result = [] if row_len > max_row_len: max_row_len = row_len for header, value in zip(padded_headers, row): if value is None: value = '<null>' row_result.append(u"%s %s" % (header, value)) results.append('\n'.join(row_result)) output = [] for i, result in enumerate(results): output.append(get_separator(i, header_len, max_row_len)) output.append(result) output.append('\n') return ''.join(output)
Make the null value consistent between vertical and tabular output.
Make the null value consistent between vertical and tabular output.
Python
bsd-3-clause
MnO2/rediscli,martijnengler/mycli,qbdsoft/mycli,j-bennet/mycli,D-e-e-m-o/mycli,mdsrosa/mycli,martijnengler/mycli,mattn/mycli,jinstrive/mycli,webwlsong/mycli,brewneaux/mycli,danieljwest/mycli,thanatoskira/mycli,ksmaheshkumar/mycli,brewneaux/mycli,MnO2/rediscli,evook/mycli,webwlsong/mycli,j-bennet/mycli,evook/mycli,suzukaze/mycli,chenpingzhao/mycli,shoma/mycli,ZuoGuocai/mycli,tkuipers/mycli,ksmaheshkumar/mycli,mattn/mycli,ZuoGuocai/mycli,douglasvegas/mycli,oguzy/mycli,nkhuyu/mycli,oguzy/mycli,douglasvegas/mycli,shoma/mycli,qbdsoft/mycli,tkuipers/mycli,danieljwest/mycli,suzukaze/mycli,mdsrosa/mycli,jinstrive/mycli,chenpingzhao/mycli,nkhuyu/mycli,D-e-e-m-o/mycli,thanatoskira/mycli
--- +++ @@ -23,6 +23,7 @@ max_row_len = row_len for header, value in zip(padded_headers, row): + if value is None: value = '<null>' row_result.append(u"%s %s" % (header, value)) results.append('\n'.join(row_result))
495a368098357c94ccec2f9f84c077fd5b27bde7
tests/PushoverAPI/test_PushoverAPI.py
tests/PushoverAPI/test_PushoverAPI.py
from urllib.parse import urljoin, parse_qs import responses from tests.constants import * from tests.fixtures import PushoverAPI from tests.util import messages_callback @responses.activate def test_PushoverAPI_sends_message(PushoverAPI): responses.add_callback( responses.POST, urljoin(PUSHOVER_API_URL, 'messages.json'), callback=messages_callback, content_type='application/json' ) resp = PushoverAPI.send_message(TEST_USER, TEST_MESSAGE) request_body = parse_qs(resp.request.body) assert request_body['token'][0] == TEST_TOKEN assert request_body['user'][0] == TEST_USER assert request_body['message'][0] == TEST_MESSAGE assert request_body['html'][0] == 'False'
from urllib.parse import urljoin, parse_qs import responses from tests.constants import * from tests.fixtures import PushoverAPI from tests.util import messages_callback @responses.activate def test_PushoverAPI_sends_message(PushoverAPI): responses.add_callback( responses.POST, urljoin(PUSHOVER_API_URL, 'messages.json'), callback=messages_callback, content_type='application/json' ) resp = PushoverAPI.send_message(TEST_USER, TEST_MESSAGE) request_body = parse_qs(resp.request.body) assert request_body['token'][0] == TEST_TOKEN assert request_body['user'][0] == TEST_USER assert request_body['message'][0] == TEST_MESSAGE assert request_body['html'][0] == 'False' assert resp.json() == { 'status': 1, 'request': TEST_REQUEST_ID }
Add a test for a proper response, not just proper request
Add a test for a proper response, not just proper request
Python
mit
scolby33/pushover_complete
--- +++ @@ -22,3 +22,8 @@ assert request_body['message'][0] == TEST_MESSAGE assert request_body['html'][0] == 'False' + assert resp.json() == { + 'status': 1, + 'request': TEST_REQUEST_ID + } +
dc86283bb517c56eec177804801f66227477a097
networkx/generators/ego.py
networkx/generators/ego.py
""" Ego graph. """ # Copyright (C) 2010 by # Aric Hagberg <hagberg@lanl.gov> # Dan Schult <dschult@colgate.edu> # Pieter Swart <swart@lanl.gov> # All rights reserved. # BSD license. __author__ = """\n""".join(['Drew Conway <drew.conway@nyu.edu>', 'Aric Hagberg <hagberg@lanl.gov>']) __all__ = ['ego_graph'] import networkx as nx def ego_graph(G,n,radius=1,center=True): """Returns induced subgraph of neighbors centered at node n. Parameters ---------- G : graph A NetworkX Graph or DiGraph n : node A single node radius : integer Include all neighbors of distance<=radius from n center : bool, optional If False, do not include center node in graph """ sp=nx.single_source_shortest_path_length(G,n,cutoff=radius) H=G.subgraph(sp.keys()) if not center: H.remove_node(n) return H
""" Ego graph. """ # Copyright (C) 2010 by # Aric Hagberg <hagberg@lanl.gov> # Dan Schult <dschult@colgate.edu> # Pieter Swart <swart@lanl.gov> # All rights reserved. # BSD license. __author__ = """\n""".join(['Drew Conway <drew.conway@nyu.edu>', 'Aric Hagberg <hagberg@lanl.gov>']) __all__ = ['ego_graph'] import networkx as nx def ego_graph(G,n,radius=1,center=True): """Returns induced subgraph of neighbors centered at node n. Parameters ---------- G : graph A NetworkX Graph or DiGraph n : node A single node radius : integer Include all neighbors of distance<=radius from n center : bool, optional If False, do not include center node in graph Notes ----- For directed graphs D this produces the "out" neighborhood or successors. If you want the neighborhood of predecessors first reverse the graph with D.reverse(). If you want both first convert the graph to an undirected graph using G=nx.Graph(D). """ sp=nx.single_source_shortest_path_length(G,n,cutoff=radius) H=G.subgraph(sp.keys()) if not center: H.remove_node(n) return H
Add note about directed graphs
Add note about directed graphs --HG-- extra : convert_revision : svn%3A3ed01bd8-26fb-0310-9e4c-ca1a4053419f/networkx/trunk%401776
Python
bsd-3-clause
ghdk/networkx,farhaanbukhsh/networkx,yashu-seth/networkx,bzero/networkx,nathania/networkx,jakevdp/networkx,goulu/networkx,Sixshaman/networkx,ionanrozenfeld/networkx,chrisnatali/networkx,jcurbelo/networkx,tmilicic/networkx,chrisnatali/networkx,RMKD/networkx,ionanrozenfeld/networkx,sharifulgeo/networkx,sharifulgeo/networkx,ghdk/networkx,jtorrents/networkx,kai5263499/networkx,nathania/networkx,aureooms/networkx,sharifulgeo/networkx,RMKD/networkx,OrkoHunter/networkx,dhimmel/networkx,nathania/networkx,bzero/networkx,kai5263499/networkx,jfinkels/networkx,dmoliveira/networkx,kernc/networkx,dmoliveira/networkx,ltiao/networkx,kai5263499/networkx,cmtm/networkx,debsankha/networkx,dhimmel/networkx,beni55/networkx,RMKD/networkx,chrisnatali/networkx,harlowja/networkx,ionanrozenfeld/networkx,JamesClough/networkx,blublud/networkx,dhimmel/networkx,aureooms/networkx,blublud/networkx,farhaanbukhsh/networkx,ghdk/networkx,dmoliveira/networkx,harlowja/networkx,jni/networkx,wasade/networkx,blublud/networkx,SanketDG/networkx,debsankha/networkx,jni/networkx,jakevdp/networkx,debsankha/networkx,farhaanbukhsh/networkx,michaelpacer/networkx,bzero/networkx,andnovar/networkx,jtorrents/networkx,jni/networkx,aureooms/networkx,jakevdp/networkx,kernc/networkx,harlowja/networkx,NvanAdrichem/networkx,kernc/networkx
--- +++ @@ -31,6 +31,13 @@ center : bool, optional If False, do not include center node in graph + Notes + ----- + For directed graphs D this produces the "out" neighborhood + or successors. If you want the neighborhood of predecessors + first reverse the graph with D.reverse(). If you want both + first convert the graph to an undirected graph using G=nx.Graph(D). + """ sp=nx.single_source_shortest_path_length(G,n,cutoff=radius) H=G.subgraph(sp.keys())
21f7d85d5f22834e04a25ea23eabfd07b279bfe6
openedx/features/badging/constants.py
openedx/features/badging/constants.py
CONVERSATIONALIST = ('conversationalist', 'Conversationalist') TEAM_PLAYER = ('team', 'Team player') BADGES_KEY = 'badges' BADGE_NOT_FOUND_ERROR = 'There exists no badge with id {badge_id}' BADGE_TYPE_ERROR = 'Cannot assign badge {badge_id} of unknown type {badge_type}' FILTER_BADGES_ERROR = 'Unable to get badges for team {team_id}' INVALID_COMMUNITY_ERROR = 'Cannot assign badge {badge_id} for invalid community {community_id}' INVALID_TEAM_ERROR = 'Cannot assign badge {badge_id} for invalid team {community_id}' TEAM_ID_KEY = 'team_id' TEAM_ROOM_ID_KEY = 'team__room_id' UNKNOWN_COURSE_ERROR = 'Cannot assign badge {badge_id} for team {community_id} in unknown course'
CONVERSATIONALIST = ('conversationalist', 'Conversationalist') TEAM_PLAYER = ('team', 'Team player') BADGES_KEY = 'badges' BADGE_NOT_FOUND_ERROR = 'There exists no badge with id {badge_id}' BADGE_TYPE_ERROR = 'Cannot assign badge {badge_id} of unknown type {badge_type}' BADGE_ROOT_URL = '{root_url}/courses/{course_id}' FILTER_BADGES_ERROR = 'Unable to get badges for team {team_id}' INVALID_COMMUNITY_ERROR = 'Cannot assign badge {badge_id} for invalid community {community_id}' INVALID_TEAM_ERROR = 'Cannot assign badge {badge_id} for invalid team {community_id}' TEAM_ID_KEY = 'team_id' TEAM_ROOM_ID_KEY = 'team__room_id' UNKNOWN_COURSE_ERROR = 'Cannot assign badge {badge_id} for team {community_id} in unknown course'
Add constant for badge url
Add constant for badge url
Python
agpl-3.0
philanthropy-u/edx-platform,philanthropy-u/edx-platform,philanthropy-u/edx-platform,philanthropy-u/edx-platform
--- +++ @@ -4,6 +4,7 @@ BADGES_KEY = 'badges' BADGE_NOT_FOUND_ERROR = 'There exists no badge with id {badge_id}' BADGE_TYPE_ERROR = 'Cannot assign badge {badge_id} of unknown type {badge_type}' +BADGE_ROOT_URL = '{root_url}/courses/{course_id}' FILTER_BADGES_ERROR = 'Unable to get badges for team {team_id}' INVALID_COMMUNITY_ERROR = 'Cannot assign badge {badge_id} for invalid community {community_id}' INVALID_TEAM_ERROR = 'Cannot assign badge {badge_id} for invalid team {community_id}'
3b5880c375ee92ce931c29a978ff64cd8849d028
src/simple-http-server.py
src/simple-http-server.py
#!/usr/bin/env python3 from argparse import ArgumentParser, ArgumentDefaultsHelpFormatter import os, http.server def main(args): os.chdir(args.directory) addr = ('' ,args.port) httpd = http.server.HTTPServer(addr, http.server.SimpleHTTPRequestHandler) httpd.serve_forever() if __name__ == '__main__': parser = ArgumentParser(formatter_class=ArgumentDefaultsHelpFormatter) parser.add_argument('-p', '--port', default=8000, help='Port on which to listen') parser.add_argument('-d', '--directory', metavar='DIR', default=os.getcwd(), help='Directory to serve') args = parser.parse_args() try: exit(main(args)) except KeyboardInterrupt as e: pass
#!/usr/bin/env python3 from argparse import ArgumentParser, ArgumentDefaultsHelpFormatter import os, http.server def main(args): os.chdir(args.directory) addr = ('' ,args.port) print('Serving', args.directory, 'on', addr) httpd = http.server.HTTPServer(addr, http.server.SimpleHTTPRequestHandler) httpd.serve_forever() if __name__ == '__main__': parser = ArgumentParser(formatter_class=ArgumentDefaultsHelpFormatter) parser.add_argument('-p', '--port', default=8000, help='Port on which to listen') parser.add_argument('-d', '--directory', metavar='DIR', default=os.getcwd(), help='Directory to serve') args = parser.parse_args() try: exit(main(args)) except KeyboardInterrupt as e: pass
Add message about where we're listening for connections
Add message about where we're listening for connections
Python
unlicense
pastly/python-snippits
--- +++ @@ -5,6 +5,7 @@ def main(args): os.chdir(args.directory) addr = ('' ,args.port) + print('Serving', args.directory, 'on', addr) httpd = http.server.HTTPServer(addr, http.server.SimpleHTTPRequestHandler) httpd.serve_forever()
22382935be99e027da46303107926a15cd8f3017
tests/twisted/vcard/test-set-alias.py
tests/twisted/vcard/test-set-alias.py
""" Test alias setting support. """ from servicetest import EventPattern from gabbletest import exec_test, acknowledge_iq import constants as cs def test(q, bus, conn, stream): iq_event = q.expect('stream-iq', to=None, query_ns='vcard-temp', query_name='vCard') acknowledge_iq(stream, iq_event.stanza) conn.Aliasing.SetAliases({1: 'lala'}) iq_event = q.expect('stream-iq', iq_type='set', query_ns='vcard-temp', query_name='vCard') acknowledge_iq(stream, iq_event.stanza) event = q.expect('dbus-signal', signal='AliasesChanged', args=[[(1, u'lala')]]) if __name__ == '__main__': exec_test(test)
""" Test alias setting support. """ from servicetest import EventPattern, assertEquals from gabbletest import exec_test, acknowledge_iq import constants as cs import ns def validate_pep_update(pep_update, expected_nickname): publish = pep_update.query.elements(uri=ns.PUBSUB, name='publish').next() assertEquals(ns.NICK, publish['node']) item = publish.elements(uri=ns.PUBSUB, name='item').next() nick = item.elements(uri=ns.NICK, name='nick').next() assertEquals(expected_nickname, nick.children[0]) def test(q, bus, conn, stream): iq_event = q.expect('stream-iq', to=None, query_ns='vcard-temp', query_name='vCard') acknowledge_iq(stream, iq_event.stanza) conn.Aliasing.SetAliases({1: 'lala'}) pep_update = q.expect('stream-iq', iq_type='set', query_ns=ns.PUBSUB, query_name='pubsub') validate_pep_update(pep_update, 'lala') acknowledge_iq(stream, pep_update.stanza) iq_event = q.expect('stream-iq', iq_type='set', query_ns='vcard-temp', query_name='vCard') acknowledge_iq(stream, iq_event.stanza) event = q.expect('dbus-signal', signal='AliasesChanged', args=[[(1, u'lala')]]) if __name__ == '__main__': exec_test(test)
Test setting our own alias via PEP
Test setting our own alias via PEP Astonishingly, this was untested...
Python
lgpl-2.1
Ziemin/telepathy-gabble,jku/telepathy-gabble,mlundblad/telepathy-gabble,Ziemin/telepathy-gabble,mlundblad/telepathy-gabble,jku/telepathy-gabble,Ziemin/telepathy-gabble,mlundblad/telepathy-gabble,jku/telepathy-gabble,Ziemin/telepathy-gabble
--- +++ @@ -3,9 +3,17 @@ Test alias setting support. """ -from servicetest import EventPattern +from servicetest import EventPattern, assertEquals from gabbletest import exec_test, acknowledge_iq import constants as cs +import ns + +def validate_pep_update(pep_update, expected_nickname): + publish = pep_update.query.elements(uri=ns.PUBSUB, name='publish').next() + assertEquals(ns.NICK, publish['node']) + item = publish.elements(uri=ns.PUBSUB, name='item').next() + nick = item.elements(uri=ns.NICK, name='nick').next() + assertEquals(expected_nickname, nick.children[0]) def test(q, bus, conn, stream): iq_event = q.expect('stream-iq', to=None, query_ns='vcard-temp', @@ -14,6 +22,10 @@ acknowledge_iq(stream, iq_event.stanza) conn.Aliasing.SetAliases({1: 'lala'}) + + pep_update = q.expect('stream-iq', iq_type='set', query_ns=ns.PUBSUB, query_name='pubsub') + validate_pep_update(pep_update, 'lala') + acknowledge_iq(stream, pep_update.stanza) iq_event = q.expect('stream-iq', iq_type='set', query_ns='vcard-temp', query_name='vCard')
fe9f48415017bcd873140da82a5f2e463d13b307
tests/scoring_engine/models/test_setting.py
tests/scoring_engine/models/test_setting.py
from scoring_engine.models.setting import Setting from tests.scoring_engine.unit_test import UnitTest class TestSetting(UnitTest): def test_init_setting(self): setting = Setting(name='test_setting', value='test value example') assert setting.id is None assert setting.name == 'test_setting' assert setting.value == 'test value example' self.session.add(setting) self.session.commit() assert setting.id is not None def test_get_setting(self): setting_old = Setting(name='test_setting', value='test value example') self.session.add(setting_old) setting_new = Setting(name='test_setting', value='updated example') self.session.add(setting_new) self.session.commit() assert Setting.get_setting('test_setting').value == 'updated example' def test_boolean_value(self): setting = Setting(name='test_setting', value=True) assert setting.name == 'test_setting' assert setting.value is True self.session.add(setting) self.session.commit() assert setting.value is True
from scoring_engine.models.setting import Setting from tests.scoring_engine.unit_test import UnitTest class TestSetting(UnitTest): def test_init_setting(self): setting = Setting(name='test_setting', value='test value example') assert setting.id is None assert setting.name == 'test_setting' assert setting.value == 'test value example' assert setting._value_type == 'String' self.session.add(setting) self.session.commit() assert setting.id is not None def test_get_setting(self): setting_old = Setting(name='test_setting', value='test value example') self.session.add(setting_old) setting_new = Setting(name='test_setting', value='updated example') self.session.add(setting_new) self.session.commit() assert Setting.get_setting('test_setting').value == 'updated example' def test_boolean_value(self): setting = Setting(name='test_setting', value=True) assert setting.name == 'test_setting' assert setting.value is True self.session.add(setting) self.session.commit() assert setting.value is True def test_boolean_value_advanced(self): setting = Setting(name='test_setting', value=True) assert setting.name == 'test_setting' assert setting.value is True self.session.add(setting) self.session.commit() setting.value = 'somevalue' assert setting.value == 'somevalue' self.session.add(setting) self.session.commit()
Add test for setting as boolean value
Add test for setting as boolean value
Python
mit
pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine
--- +++ @@ -10,6 +10,7 @@ assert setting.id is None assert setting.name == 'test_setting' assert setting.value == 'test value example' + assert setting._value_type == 'String' self.session.add(setting) self.session.commit() assert setting.id is not None @@ -29,3 +30,14 @@ self.session.add(setting) self.session.commit() assert setting.value is True + + def test_boolean_value_advanced(self): + setting = Setting(name='test_setting', value=True) + assert setting.name == 'test_setting' + assert setting.value is True + self.session.add(setting) + self.session.commit() + setting.value = 'somevalue' + assert setting.value == 'somevalue' + self.session.add(setting) + self.session.commit()
bd5336d52a1cd2e086dbf1665b90823505d31840
tests/testapp/tests/test_timer_collector.py
tests/testapp/tests/test_timer_collector.py
from datetime import timedelta from django_performance_testing.timing import TimeCollector from freezegun import freeze_time import pytest from testapp.test_helpers import capture_result_collected @pytest.mark.parametrize('seconds', [10, 5, 0.04]) def test_captures_and_measures_elapsed_time(seconds): with capture_result_collected() as captured: with freeze_time('2016-09-22 15:57:01') as frozen_time: with TimeCollector(): frozen_time.tick(timedelta(seconds=seconds)) assert len(captured.calls) == 1 assert pytest.approx(seconds) == captured.calls[0]['results']
from datetime import timedelta from django_performance_testing.timing import TimeCollector from freezegun import freeze_time import pytest from testapp.test_helpers import capture_result_collected @pytest.mark.parametrize('seconds', [10, 5, 0.04]) def test_captures_and_measures_elapsed_time(seconds): with capture_result_collected() as captured: with freeze_time('2016-09-22 15:57:01') as frozen_time: with TimeCollector(): frozen_time.tick(timedelta(seconds=seconds)) assert len(captured.calls) == 1 assert pytest.approx(seconds) == captured.calls[0]['results'][0].value
Fix for failing test test_captures_and_measures_elapsed_time: TypeError: unsupported operand type(s) for -: 'float' and 'list'
Fix for failing test test_captures_and_measures_elapsed_time: TypeError: unsupported operand type(s) for -: 'float' and 'list'
Python
bsd-3-clause
PaesslerAG/django-performance-testing
--- +++ @@ -12,4 +12,4 @@ with TimeCollector(): frozen_time.tick(timedelta(seconds=seconds)) assert len(captured.calls) == 1 - assert pytest.approx(seconds) == captured.calls[0]['results'] + assert pytest.approx(seconds) == captured.calls[0]['results'][0].value
4456f5604c1d824f5012bcee550d274c905d74c8
bigcrunch/shutdown.py
bigcrunch/shutdown.py
import asyncio import botocore.exceptions from bigcrunch import webapp @asyncio.coroutine def shutdown(): client = yield from webapp.redshift_client() cluster_control = webapp.ClusterControl(client) try: cluster = yield from cluster_control.get() except botocore.exceptions.ClientError as e: if e.response['Error']['Code'] != 'ClusterNotFound': raise e else: print('Redshift already shutdown') return engine = yield from webapp.create_engine(cluster) with (yield from engine) as conn: db = webapp.Database(conn=conn) sessions = yield from db.running_test_sessions() if sessions == 0: print('shutting down Redshift') yield from cluster_control.destroy() if __name__ == '__main__': loop = asyncio.get_event_loop() loop.run_until_complete(shutdown()) loop.close()
import asyncio import botocore.exceptions from bigcrunch import webapp @asyncio.coroutine def shutdown(): client = webapp.redshift_client() cluster_control = webapp.ClusterControl(client) try: cluster = yield from cluster_control.get() except botocore.exceptions.ClientError as e: if e.response['Error']['Code'] != 'ClusterNotFound': raise e else: print('Redshift already shutdown') return engine = yield from webapp.create_engine(cluster) with (yield from engine) as conn: db = webapp.Database(conn=conn) sessions = yield from db.running_test_sessions() if sessions == 0: print('shutting down Redshift') yield from cluster_control.destroy() if __name__ == '__main__': loop = asyncio.get_event_loop() loop.run_until_complete(shutdown()) loop.close()
Remove yield from on redshift_client
Remove yield from on redshift_client
Python
agpl-3.0
sqlalchemy-redshift/bigcrunch
--- +++ @@ -7,7 +7,7 @@ @asyncio.coroutine def shutdown(): - client = yield from webapp.redshift_client() + client = webapp.redshift_client() cluster_control = webapp.ClusterControl(client) try: cluster = yield from cluster_control.get()
1c2ade08262259bb952424d7605d5fc68e038d73
scripts/generate_labos_aggrega.py
scripts/generate_labos_aggrega.py
#!/usr/bin/env python from csv import reader with open("data/public/labos.departements.csv") as f: data = list(reader(f)) output = {} keys = data.pop(0) keys.remove("LABO") keys.remove("DEPARTEMENT") for row in data: if not row[0]: continue if row[0] not in output: output[row[0]] = dict({k: 0 for k in keys}) for i, k in enumerate(keys): output[row[0]][k] += float(row[2+i]) if row[2+i] else 0 print "LABO,"+",".join(keys) for labo in sorted(output.keys(), key=lambda x: -output[x]["MONTANT AVANTAGES"]): print labo+","+",".join([str(output[labo][k]) for k in keys])
#!/usr/bin/env python from csv import reader with open("data/public/labos.departements.csv") as f: data = list(reader(f)) output = {} keys = data.pop(0) keys.remove("LABO") keys.remove("DEPARTEMENT") for row in data: if not row[0]: continue if row[0] not in output: output[row[0]] = dict({k: 0 for k in keys}) for i, k in enumerate(keys): output[row[0]][k] += float(row[2+i]) if row[2+i] else 0 print("LABO,"+",".join(keys)) for labo in sorted(output.keys(), key=lambda x: -output[x]["MONTANT AVANTAGES"]): print(labo+","+",".join([str(output[labo][k]) for k in keys]))
Make aggregation script compatible with python 3
Make aggregation script compatible with python 3
Python
agpl-3.0
regardscitoyens/sunshine-data,regardscitoyens/sunshine-data,regardscitoyens/sunshine-data,regardscitoyens/sunshine-data,regardscitoyens/sunshine-data,regardscitoyens/sunshine-data
--- +++ @@ -17,6 +17,6 @@ for i, k in enumerate(keys): output[row[0]][k] += float(row[2+i]) if row[2+i] else 0 -print "LABO,"+",".join(keys) +print("LABO,"+",".join(keys)) for labo in sorted(output.keys(), key=lambda x: -output[x]["MONTANT AVANTAGES"]): - print labo+","+",".join([str(output[labo][k]) for k in keys]) + print(labo+","+",".join([str(output[labo][k]) for k in keys]))
f25a32dd0180af91277ace186fc878c8baffed65
heisen/core/__init__.py
heisen/core/__init__.py
from heisen.config import settings from heisen.core.log import logger from jsonrpclib.request import Connection def get_rpc_connection(): servers = { 'self': [ ('127.0.0.1', settings.RPC_PORT, 'aliehsanmilad', 'Key1_s!3cr3t') ], } servers.update(getattr(settings, 'RPC_SERVERS', {})) return Connection(servers, 'heisen', settings.APP_NAME) rpc_call = get_rpc_connection()
from heisen.config import settings from heisen.core.log import logger from jsonrpclib.request import ConnectionPool def get_rpc_connection(): if settings.CREDENTIALS: username, passowrd = settings.CREDENTIALS[0] else: username = passowrd = None servers = {'self': [('localhost', settings.RPC_PORT, username, passowrd)]} servers.update(getattr(settings, 'RPC_SERVERS', {})) return ConnectionPool(servers, 'heisen', settings.APP_NAME) rpc_call = get_rpc_connection()
Use connection pool for jsonrpc
Use connection pool for jsonrpc
Python
mit
HeisenCore/heisen
--- +++ @@ -1,17 +1,18 @@ from heisen.config import settings from heisen.core.log import logger -from jsonrpclib.request import Connection +from jsonrpclib.request import ConnectionPool def get_rpc_connection(): - servers = { - 'self': [ - ('127.0.0.1', settings.RPC_PORT, 'aliehsanmilad', 'Key1_s!3cr3t') - ], - } + if settings.CREDENTIALS: + username, passowrd = settings.CREDENTIALS[0] + else: + username = passowrd = None + + servers = {'self': [('localhost', settings.RPC_PORT, username, passowrd)]} servers.update(getattr(settings, 'RPC_SERVERS', {})) - return Connection(servers, 'heisen', settings.APP_NAME) + return ConnectionPool(servers, 'heisen', settings.APP_NAME) rpc_call = get_rpc_connection()
57164efc9827f7975cdfa171a5a88e6fcc4059e5
neutron/conf/policies/service_type.py
neutron/conf/policies/service_type.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from neutron.conf.policies import base rules = [ policy.DocumentedRuleDefault( 'get_service_provider', base.RULE_ANY, 'Get service providers', [ { 'method': 'GET', 'path': '/service-providers', }, ] ), ] def list_rules(): return rules
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_log import versionutils from oslo_policy import policy from neutron.conf.policies import base DEPRECATION_REASON = ( "The Service Providers API now supports system scope and default roles.") rules = [ policy.DocumentedRuleDefault( name='get_service_provider', check_str=base.SYSTEM_OR_PROJECT_READER, description='Get service providers', operations=[ { 'method': 'GET', 'path': '/service-providers', }, ], scope_types=['system', 'project'], deprecated_rule=policy.DeprecatedRule( name='get_service_provider', check_str=base.RULE_ANY), deprecated_reason=DEPRECATION_REASON, deprecated_since=versionutils.deprecated.WALLABY ), ] def list_rules(): return rules
Implement secure RBAC for service providers API
Implement secure RBAC for service providers API This commit updates the policies for service providers API to understand scope checking and account for a read-only role. This is part of a broader series of changes across OpenStack to provide a consistent RBAC experience and improve security. Partially-Implements blueprint: secure-rbac-roles Change-Id: I76af24a5153f1bc38b630da42de07f0f7b40cc41
Python
apache-2.0
mahak/neutron,mahak/neutron,openstack/neutron,openstack/neutron,openstack/neutron,mahak/neutron
--- +++ @@ -10,22 +10,32 @@ # License for the specific language governing permissions and limitations # under the License. +from oslo_log import versionutils from oslo_policy import policy from neutron.conf.policies import base +DEPRECATION_REASON = ( + "The Service Providers API now supports system scope and default roles.") + rules = [ policy.DocumentedRuleDefault( - 'get_service_provider', - base.RULE_ANY, - 'Get service providers', - [ + name='get_service_provider', + check_str=base.SYSTEM_OR_PROJECT_READER, + description='Get service providers', + operations=[ { 'method': 'GET', 'path': '/service-providers', }, - ] + ], + scope_types=['system', 'project'], + deprecated_rule=policy.DeprecatedRule( + name='get_service_provider', + check_str=base.RULE_ANY), + deprecated_reason=DEPRECATION_REASON, + deprecated_since=versionutils.deprecated.WALLABY ), ]
c916cc05c57cdda5a9b07901af8af032a97c3841
karma/api/__init__.py
karma/api/__init__.py
#!/usr/bin/env python from flask import Blueprint from flask.ext.restful import Api api = Blueprint("api", __name__, template_folder="templates") rest = Api(api) from karma.api.charity import Charity, Charities from karma.api.post import Post, Posts # Adding resources rest.add_resource(Charity, '/charities/<int:charity_name>') rest.add_resource(Charities, '/charities') rest.add_resource(Post, '/posts/<int:post_id>') rest.add_resource(Posts, '/posts')
#!/usr/bin/env python from flask import Blueprint from flask.ext.restful import Api api = Blueprint("api", __name__, template_folder="templates") rest = Api(api) from karma.api.charity import Charity, Charities from karma.api.post import Post, Posts # Adding resources rest.add_resource(Charity, '/charities/<int:charity_name>') rest.add_resource(Charities, '/charities/') rest.add_resource(Post, '/posts/<int:post_id>') rest.add_resource(Posts, '/posts/')
Add changes to url thingies
Add changes to url thingies
Python
mit
hamhut1066/ghu-hack,hamhut1066/ghu-hack
--- +++ @@ -15,6 +15,6 @@ # Adding resources rest.add_resource(Charity, '/charities/<int:charity_name>') -rest.add_resource(Charities, '/charities') +rest.add_resource(Charities, '/charities/') rest.add_resource(Post, '/posts/<int:post_id>') -rest.add_resource(Posts, '/posts') +rest.add_resource(Posts, '/posts/')
026d887dd85bdcc5db97d65706509a18252de8ba
l10n_ar_aeroo_sale/__openerp__.py
l10n_ar_aeroo_sale/__openerp__.py
# -*- coding: utf-8 -*- { 'name': 'Argentinian Like Sale Order Aeroo Report', 'version': '1.0', 'category': 'Localization/Argentina', 'sequence': 14, 'summary': '', 'description': """ Argentinian Like Sale Order / Quotation Aeroo Report ==================================================== """, 'author': 'Ingenieria ADHOC', 'website': 'www.ingadhoc.com', 'images': [ ], 'depends': [ 'report_extended_sale', 'l10n_ar_aeroo_base', 'portal_sale', ], 'data': [ 'report_configuration_defaults_data.xml', 'sale_order_report.xml', 'sale_order_template.xml', ], 'demo': [ ], 'test': [ ], 'installable': True, 'auto_install': False, 'application': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
# -*- coding: utf-8 -*- { 'name': 'Argentinian Like Sale Order Aeroo Report', 'version': '1.0', 'category': 'Localization/Argentina', 'sequence': 14, 'summary': '', 'description': """ Argentinian Like Sale Order / Quotation Aeroo Report ==================================================== """, 'author': 'Ingenieria ADHOC', 'website': 'www.ingadhoc.com', 'images': [ ], 'depends': [ 'report_extended_sale', 'l10n_ar_aeroo_base', 'l10n_ar_aeroo_invoice', #esta dependencia es porque actualizamos algo que crea portal_sale con un valor de las invoice 'portal_sale', ], 'data': [ 'report_configuration_defaults_data.xml', 'sale_order_report.xml', 'sale_order_template.xml', ], 'demo': [ ], 'test': [ ], 'installable': True, 'auto_install': False, 'application': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
FIX dependency on aeroo rpoert
FIX dependency on aeroo rpoert
Python
agpl-3.0
ingadhoc/argentina-reporting
--- +++ @@ -16,6 +16,7 @@ 'depends': [ 'report_extended_sale', 'l10n_ar_aeroo_base', + 'l10n_ar_aeroo_invoice', #esta dependencia es porque actualizamos algo que crea portal_sale con un valor de las invoice 'portal_sale', ], 'data': [
6f745ae05a22031a36cb5cedc6b627cbf7ba6512
import_goodline_iptv.py
import_goodline_iptv.py
#!/usr/bin/env python3 import configargparse from goodline_iptv.importer import do_import if __name__ == '__main__': parser = configargparse.ArgParser() parser.add_argument('-o', '--out-dir', required=True, help='Output directory') parser.add_argument('-e', '--encoding', default='cp1251', env_var='ENCODING', help='Source JTV teleguide encoding') parser.add_argument('-t', '--timezone', default='+0700', env_var='TIMEZONE', help='Source JTV teleguide timezone') parser.add_argument('-u', '--udpxy', env_var='UDPXY_ADDR', help='Address of the udproxy service, if available') parser.add_argument('-v', '--verbosity', default=0, env_var='VERBOSITY', type=int, choices=range(0, 4), help='Verbosity level') args = parser.parse_args() do_import(args.verbosity, args.out_dir, args.encoding, args.timezone, args.udpxy)
#!/usr/bin/env python3 import configargparse from goodline_iptv.importer import do_import if __name__ == '__main__': parser = configargparse.ArgParser() parser.add_argument('-o', '--out-dir', required=True, env_var='OUTDIR', help='Output directory') parser.add_argument('-e', '--encoding', default='cp1251', env_var='ENCODING', help='Source JTV teleguide encoding') parser.add_argument('-t', '--timezone', default='+0700', env_var='TIMEZONE', help='Source JTV teleguide timezone') parser.add_argument('-u', '--udpxy', env_var='UDPXY_ADDR', help='Address of the udproxy service, if available') parser.add_argument('-v', '--verbosity', default=0, env_var='VERBOSITY', type=int, choices=range(0, 4), help='Verbosity level') args = parser.parse_args() do_import(args.verbosity, args.out_dir, args.encoding, args.timezone, args.udpxy)
Add environment variable for ability to set the output directory
Add environment variable for ability to set the output directory
Python
mit
nsadovskiy/goodline_tv
--- +++ @@ -7,7 +7,10 @@ if __name__ == '__main__': parser = configargparse.ArgParser() - parser.add_argument('-o', '--out-dir', required=True, help='Output directory') + parser.add_argument('-o', '--out-dir', + required=True, + env_var='OUTDIR', + help='Output directory') parser.add_argument('-e', '--encoding', default='cp1251',
5c8fd314dd89d8964cc5cb75c2b57ed32be275c5
plsync/users.py
plsync/users.py
# NOTE: User roles are not managed here. Visit PlanetLab to change user roles. user_list = [('Stephen', 'Stuart', 'sstuart@google.com'), ('Will', 'Hawkins', 'hawkinsw@opentechinstitute.org'), ('Jordan', 'McCarthy', 'mccarthy@opentechinstitute.org'), ('Chris', 'Ritzo', 'critzo@opentechinstitute.org'), ('Josh', 'Bailey', 'joshb@google.com'), ('Steph', 'Alarcon', 'salarcon@measurementlab.net'), ('Nathan', 'Kinkade', 'kinkade@opentechinstitute.org'), ('Matt', 'Mathis', 'mattmathis@google.com')]
# NOTE: User roles are not managed here. Visit PlanetLab to change user roles. user_list = [('Stephen', 'Stuart', 'sstuart@google.com'), ('Will', 'Hawkins', 'hawkinsw@opentechinstitute.org'), ('Jordan', 'McCarthy', 'mccarthy@opentechinstitute.org'), ('Chris', 'Ritzo', 'critzo@opentechinstitute.org'), ('Josh', 'Bailey', 'joshb@google.com'), ('Steph', 'Alarcon', 'salarcon@measurementlab.net'), ('Nathan', 'Kinkade', 'kinkade@opentechinstitute.org'), ('Matt', 'Mathis', 'mattmathis@google.com') ('Peter', 'Boothe', 'pboothe@google.com')]
Add pboothe temporarily for testing
Add pboothe temporarily for testing
Python
apache-2.0
jheretic/operator,stephen-soltesz/operator,nkinkade/operator,critzo/operator,critzo/operator,jheretic/operator,nkinkade/operator,m-lab/operator,salarcon215/operator,m-lab/operator,salarcon215/operator,stephen-soltesz/operator
--- +++ @@ -7,4 +7,5 @@ ('Josh', 'Bailey', 'joshb@google.com'), ('Steph', 'Alarcon', 'salarcon@measurementlab.net'), ('Nathan', 'Kinkade', 'kinkade@opentechinstitute.org'), - ('Matt', 'Mathis', 'mattmathis@google.com')] + ('Matt', 'Mathis', 'mattmathis@google.com') + ('Peter', 'Boothe', 'pboothe@google.com')]
4f9b01dcf194c5ce05640852cf8f436fa4e3bf80
manager/urls.py
manager/urls.py
from django.conf.urls import url from manager import views _PACKAGE_NAME_REGEX = r'(?P<package_name>[a-zA-Z0-9_+-]+)' _PACKAGE_NAME_REGEX_WITH_EMPTY = r'(?P<package_name>[a-zA-Z0-9_+-]*)' _BUILD_NUMBER_REGEX = r'(?P<build_number>\d+)' _REGEX = {'name': _PACKAGE_NAME_REGEX, 'number': _BUILD_NUMBER_REGEX} urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^register/$', views.package_register, name='package_register'), url(r'^register/{name}$'.format(name=_PACKAGE_NAME_REGEX_WITH_EMPTY), views.package_register_detail, name='package_register_detail'), url(r'^{name}/$'.format(**_REGEX), views.package_detail, name='package_detail'), url(r'^{name}/{number}/$'.format(**_REGEX), views.build_detail, name='build_detail'), url(r'^{name}/{number}/download/$'.format(**_REGEX), views.build_download, name='build_download'), url(r'^{name}/{number}/log/$'.format(**_REGEX), views.build_log, name='build_log'), url(r'^repository/(?P<file_name>.*)$', views.repository, name='repository'), ]
from django.conf.urls import url from manager import views _PACKAGE_NAME_REGEX = r'(?P<package_name>[a-zA-Z0-9_+-.]+)' _PACKAGE_NAME_REGEX_WITH_EMPTY = r'(?P<package_name>[a-zA-Z0-9_+-.]*)' _BUILD_NUMBER_REGEX = r'(?P<build_number>\d+)' _REGEX = {'name': _PACKAGE_NAME_REGEX, 'number': _BUILD_NUMBER_REGEX} urlpatterns = [ url(r'^$', views.package_list, name='package_list'), url(r'^register/$', views.package_register, name='package_register'), url(r'^register/{name}$'.format(name=_PACKAGE_NAME_REGEX_WITH_EMPTY), views.package_register_detail, name='package_register_detail'), url(r'^{name}/$'.format(**_REGEX), views.package_detail, name='package_detail'), url(r'^{name}/{number}/$'.format(**_REGEX), views.build_detail, name='build_detail'), url(r'^{name}/{number}/download/$'.format(**_REGEX), views.build_download, name='build_download'), url(r'^{name}/{number}/log/$'.format(**_REGEX), views.build_log, name='build_log'), url(r'^repository/(?P<file_name>.*)$', views.repository, name='repository'), ]
Allow '.' in package name
Allow '.' in package name
Python
mit
colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager,colajam93/aurpackager
--- +++ @@ -2,8 +2,8 @@ from manager import views -_PACKAGE_NAME_REGEX = r'(?P<package_name>[a-zA-Z0-9_+-]+)' -_PACKAGE_NAME_REGEX_WITH_EMPTY = r'(?P<package_name>[a-zA-Z0-9_+-]*)' +_PACKAGE_NAME_REGEX = r'(?P<package_name>[a-zA-Z0-9_+-.]+)' +_PACKAGE_NAME_REGEX_WITH_EMPTY = r'(?P<package_name>[a-zA-Z0-9_+-.]*)' _BUILD_NUMBER_REGEX = r'(?P<build_number>\d+)' _REGEX = {'name': _PACKAGE_NAME_REGEX, 'number': _BUILD_NUMBER_REGEX}
88b81ee89800592bccf0a714ae79be74507c8f29
test_engine.py
test_engine.py
import engine VALID_COORDS = [(x, y) for x in xrange(97, 105) for y in xrange(49, 57)] INVALID_COORDS = [ (0, 0), (-1, -1), (96, 49), (96, 48), (105, 49), (104, 48), (96, 56), (97, 57), (105, 56), (104, 57) ] VALID_A1 = [chr(x) + chr(y) for x in xrange(97, 105) for y in xrange(49, 57)] INVALID_A1 = ['a0', 'a9', 'h0', 'h9', 'z1', 'z8'] def test_coord_to_a1(): for coord in VALID_COORDS: assert engine._coord_to_a1.get(coord, False) is not False for coord in INVALID_COORDS: print coord assert engine._coord_to_a1.get(coord, False) is False
import engine VALID_COORDS = [(x, y) for x in xrange(97, 105) for y in xrange(49, 57)] INVALID_COORDS = [ (0, 0), (-1, -1), (96, 49), (96, 48), (105, 49), (104, 48), (96, 56), (97, 57), (105, 56), (104, 57) ] VALID_A1 = [chr(x) + chr(y) for x in xrange(97, 105) for y in xrange(49, 57)] INVALID_A1 = ['a0', 'a9', 'h0', 'h9', 'z1', 'z8'] def test_coord_to_a1(): for coord in VALID_COORDS: assert engine._coord_to_a1.get(coord, False) is not False for coord in INVALID_COORDS: assert engine._coord_to_a1.get(coord, False) is False def test_a1_to_coord(): for a1 in VALID_A1: assert engine._a1_to_coord.get(a1, False) is not False for a1 in INVALID_A1: assert engine._a1_to_coord.get(a1, False) is False
Add test_a1_to_coord() to assert that only valid board coordinates are in teh _a1_to_coord dictionary
Add test_a1_to_coord() to assert that only valid board coordinates are in teh _a1_to_coord dictionary
Python
mit
EyuelAbebe/gamer,EyuelAbebe/gamer
--- +++ @@ -17,5 +17,11 @@ for coord in VALID_COORDS: assert engine._coord_to_a1.get(coord, False) is not False for coord in INVALID_COORDS: - print coord assert engine._coord_to_a1.get(coord, False) is False + + +def test_a1_to_coord(): + for a1 in VALID_A1: + assert engine._a1_to_coord.get(a1, False) is not False + for a1 in INVALID_A1: + assert engine._a1_to_coord.get(a1, False) is False
4c3dd0c9d27af0f186f81c4fed0003a9190b4d9e
jal_stats/stats/serializers.py
jal_stats/stats/serializers.py
# from django.contrib.auth.models import User from rest_framework import serializers from .models import Activity, Stat class StatSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Stat fields = ('id', 'activity', 'reps', 'date') def create(self, validated_data): validated_data['activity'] = self.context['activity'] stat = Stat.objects.create(**validated_data) return stat class ActivitySerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Activity fields = ('id', 'full_description', 'units', 'url') class ActivityListSerializer(ActivitySerializer): stats = StatSerializer(many=True, read_only=True) class Meta: model = Activity fields = tuple(list(ActivitySerializer.Meta.fields) + ['stats']) # class UserSerializer(serializers.HyperlinkedModelSerializer): # # class Meta: # model = User # fields = ('id', 'username', 'email', 'activities')
# from django.contrib.auth.models import User from rest_framework import serializers from .models import Activity, Stat class StatAddSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Stat fields = ('id', 'reps', 'date') class StatSerializer(StatAddSerializer): class Meta: model = Stat fields = tuple(list(StatAddSerializer.Meta.fields) + ['activity']) def create(self, validated_data): validated_data['activity'] = self.context['activity'] stat = Stat.objects.create(**validated_data) return stat class ActivitySerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Activity fields = ('id', 'full_description', 'units', 'url') class ActivityListSerializer(ActivitySerializer): stats = StatSerializer(many=True, read_only=True) class Meta: model = Activity fields = tuple(list(ActivitySerializer.Meta.fields) + ['stats']) # class UserSerializer(serializers.HyperlinkedModelSerializer): # # class Meta: # model = User # fields = ('id', 'username', 'email', 'activities')
Add new serializer for StatAdd that doesn't have activity
Add new serializer for StatAdd that doesn't have activity
Python
mit
jal-stats/django
--- +++ @@ -3,11 +3,18 @@ from .models import Activity, Stat -class StatSerializer(serializers.HyperlinkedModelSerializer): +class StatAddSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = Stat - fields = ('id', 'activity', 'reps', 'date') + fields = ('id', 'reps', 'date') + + +class StatSerializer(StatAddSerializer): + + class Meta: + model = Stat + fields = tuple(list(StatAddSerializer.Meta.fields) + ['activity']) def create(self, validated_data): validated_data['activity'] = self.context['activity']
fd6702fbb43eb4e6c5129ac6026908946f03c1a7
paws/handler.py
paws/handler.py
from .request import Request from .response import response class Handler(object): ''' Simple dispatcher class. ''' def __call__(self, event, context): self.request = request = Request(event, context) func = getattr(self, self.event['httpMethod'], self.invalid) return func(request, *self.event['pathParameters']) def invalid(self, *args): return response(status=405)
from .request import Request from .response import response class Handler(object): ''' Simple dispatcher class. ''' def __init__(self, event, context): self.request = Request(event, context) def __call__(self, event, context): func = getattr(self, self.event['httpMethod'], self.invalid) return func(self.request, *self.event['pathParameters']) def invalid(self, *args): return response(status=405)
Move request construction to init
Move request construction to init
Python
bsd-3-clause
funkybob/paws
--- +++ @@ -6,11 +6,12 @@ ''' Simple dispatcher class. ''' + def __init__(self, event, context): + self.request = Request(event, context) def __call__(self, event, context): - self.request = request = Request(event, context) func = getattr(self, self.event['httpMethod'], self.invalid) - return func(request, *self.event['pathParameters']) + return func(self.request, *self.event['pathParameters']) def invalid(self, *args): return response(status=405)
513709e123c63e1eb30065b6f35c18faceac820b
user.py
user.py
import re class User: Groups = {' ':0,'+':1,'☆':1,'%':2,'@':3,'*':3.1,'&':4,'#':5,'~':6} @staticmethod def compareRanks(rank1, rank2): try: return User.Groups[rank1] >= User.Groups[rank2] except: if not rank1 in User.Groups: print('{rank} is not a supported usergroup'.format(rank = rank1)) if not rank2 in User.Groups: print('{rank} is not a supported usergroup'.format(rank = rank2)) return False def __init__(self, name, rank, owner = False): self.name = name self.id = re.sub(r'[^a-zA-z0-9]', '', name).lower() self.rank = rank self.owner = owner def hasRank(self, rank): return self.owner or User.compareRanks(self.rank, rank) def isOwner(self): return self.owner
import re class User: Groups = {' ':0,'+':1,'☆':1,'%':2,'@':3,'*':3.1,'&':4,'#':5,'~':6} @staticmethod def compareRanks(rank1, rank2): try: return User.Groups[rank1] >= User.Groups[rank2] except: if not rank1 in User.Groups: print('{rank} is not a supported usergroup'.format(rank = rank1)) if not rank2 in User.Groups: print('{rank} is not a supported usergroup'.format(rank = rank2)) return False def __init__(self, name, rank = ' ', owner = False): self.name = name self.id = re.sub(r'[^a-zA-z0-9]', '', name).lower() self.rank = rank self.owner = owner def hasRank(self, rank): return self.owner or User.compareRanks(self.rank, rank) def isOwner(self): return self.owner
Add default rank to User for simplified tests
Add default rank to User for simplified tests
Python
mit
QuiteQuiet/PokemonShowdownBot
--- +++ @@ -12,7 +12,7 @@ print('{rank} is not a supported usergroup'.format(rank = rank2)) return False - def __init__(self, name, rank, owner = False): + def __init__(self, name, rank = ' ', owner = False): self.name = name self.id = re.sub(r'[^a-zA-z0-9]', '', name).lower() self.rank = rank
bdcf90a0fdf782b1c6cfd261e0dbb208e013eb1b
python/day12.py
python/day12.py
#!/usr/local/bin/python3 import json import pathlib input_file = pathlib.Path(__file__).parent.parent.joinpath('day12_input.txt') def sum_data(d): total = 0 if isinstance(d, dict): d = d.values() for item in d: if isinstance(item, int): total += item elif isinstance(item, (list, dict)): total += sum_data(item) else: continue # Some other type we’re not interested in return total def sum_json(raw_json): parsed = json.loads(raw_json) return sum_data(parsed) def test_simple(): assert sum_json('[1,2,3]') == 6 assert sum_json('{"a":2,"b":4}') == 6 def test_nested(): assert sum_json('[[[3]]]') == 3 assert sum_json('{"a":{"b":4},"c":-1}') == 3 def test_mixed(): assert sum_json('{"a":[-1,1]}') == 0 assert sum_json('[-1,{"a":1}]') == 0 def test_empty(): assert sum_json('[]') == 0 assert sum_json('{}') == 0 if __name__ == '__main__': with open(input_file) as json_file: json_data = json_file.read() print(sum_json(json_data))
#!/usr/local/bin/python3 import json import pathlib input_file = pathlib.Path(__file__).parent.parent.joinpath('day12_input.txt') def sum_data(d): total = 0 if isinstance(d, dict): d = d.values() if 'red' in d: return 0 for item in d: if isinstance(item, int): total += item elif isinstance(item, (list, dict)): total += sum_data(item) else: continue # Some other type we’re not interested in return total def sum_json(raw_json): parsed = json.loads(raw_json) return sum_data(parsed) def test_simple(): assert sum_json('[1,2,3]') == 6 assert sum_json('{"a":2,"b":4}') == 6 def test_nested(): assert sum_json('[[[3]]]') == 3 assert sum_json('{"a":{"b":4},"c":-1}') == 3 def test_mixed(): assert sum_json('{"a":[-1,1]}') == 0 assert sum_json('[-1,{"a":1}]') == 0 def test_empty(): assert sum_json('[]') == 0 assert sum_json('{}') == 0 if __name__ == '__main__': with open(input_file) as json_file: json_data = json_file.read() print(sum_json(json_data))
Add day 12 part two solution in python
Add day 12 part two solution in python
Python
mit
robjwells/adventofcode-solutions,robjwells/adventofcode-solutions,robjwells/adventofcode-solutions,robjwells/adventofcode-solutions,robjwells/adventofcode-solutions,robjwells/adventofcode-solutions
--- +++ @@ -11,6 +11,7 @@ total = 0 if isinstance(d, dict): d = d.values() + if 'red' in d: return 0 for item in d: if isinstance(item, int): total += item
025b8915306670ac45f200ac4cbf32115eab53c7
python/setup.py
python/setup.py
from setuptools import setup # This is the cabb_scheduler Python library. # Jamie Stevens 2017 # ATCA Senior Systems Scientist # Jamie.Stevens@csiro.au setup(name='cabb_scheduler', version='1.1', description='CABB Scheduling Python Library', url='https://github.com/ste616/cabb-schedule-api', author='Jamie Stevens', author_email='Jamie.Stevens@csiro.au', license='MIT', packages=[ 'cabb_scheduler' ], install_requires=[ 'numpy', 'requests' ], zip_safe=False) # Changelog: # 2017-03-14, v1.1: Added the parse routine to the schedule library, to read/write in # strings instead of just files, and return number of scans read in.
from setuptools import setup # This is the cabb_scheduler Python library. # Jamie Stevens 2017 # ATCA Senior Systems Scientist # Jamie.Stevens@csiro.au setup(name='cabb_scheduler', version='1.2', description='CABB Scheduling Python Library', url='https://github.com/ste616/cabb-schedule-api', author='Jamie Stevens', author_email='Jamie.Stevens@csiro.au', license='MIT', packages=[ 'cabb_scheduler' ], install_requires=[ 'numpy', 'requests' ], zip_safe=False) # Changelog: # 2017-03-14, v1.1: Added the parse routine to the schedule library, to read/write in # strings instead of just files, and return number of scans read in. # 2017-03-15, v1.2: Add the ability to query MoniCA for the array name and the # frequencies currently in use.
Change the library version number.
Change the library version number.
Python
mit
ste616/cabb-schedule-api
--- +++ @@ -6,7 +6,7 @@ # Jamie.Stevens@csiro.au setup(name='cabb_scheduler', - version='1.1', + version='1.2', description='CABB Scheduling Python Library', url='https://github.com/ste616/cabb-schedule-api', author='Jamie Stevens', @@ -22,3 +22,5 @@ # Changelog: # 2017-03-14, v1.1: Added the parse routine to the schedule library, to read/write in # strings instead of just files, and return number of scans read in. +# 2017-03-15, v1.2: Add the ability to query MoniCA for the array name and the +# frequencies currently in use.
3436b94a4c69b843c65f6ddf6756c18ca540c090
linked-list/is-list-palindrome.py
linked-list/is-list-palindrome.py
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if not l.value or not l.next.value: return True def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next
# Given a singly linked list of integers, determine whether or not it's a palindrome class Node(object): # define constructor def __init__(self, value): self.value = value self.next = None def is_list_palindrome(l): if not l.value or not l.next.value: return True fake_head = Node(None) fake_head.next = l fast_node = fake_head slow_node = fake_head while fast_node.next and fast_node.next.next: fast_node = fast_node.next.next slow_node = slow_node.next def create_nodes(l): root = Node(-1) current_node = root for value in l: current_node.next = Node(value) current_node = current_node.next return root.next is_list_palindrome(create_nodes([1, 2, 3, 4]))
Create fast node that is twice the speed of slow node to get to center of list
Create fast node that is twice the speed of slow node to get to center of list
Python
mit
derekmpham/interview-prep,derekmpham/interview-prep
--- +++ @@ -8,6 +8,15 @@ if not l.value or not l.next.value: return True + fake_head = Node(None) + fake_head.next = l + fast_node = fake_head + slow_node = fake_head + + while fast_node.next and fast_node.next.next: + fast_node = fast_node.next.next + slow_node = slow_node.next + def create_nodes(l): root = Node(-1) current_node = root @@ -15,3 +24,6 @@ current_node.next = Node(value) current_node = current_node.next return root.next + + +is_list_palindrome(create_nodes([1, 2, 3, 4]))
a1a9ab2b2c0ca7749984f5ad6e6430980e0d0ecc
tests/basics/list_sort.py
tests/basics/list_sort.py
l = [1, 3, 2, 5] print(l) l.sort() print(l) l.sort(key=lambda x: -x) print(l) l.sort(key=lambda x: -x, reverse=True) print(l) l.sort(reverse=True) print(l) l.sort(reverse=False) print(l)
l = [1, 3, 2, 5] print(l) print(sorted(l)) l.sort() print(l) print(l == sorted(l)) print(sorted(l, key=lambda x: -x)) l.sort(key=lambda x: -x) print(l) print(l == sorted(l, key=lambda x: -x)) print(sorted(l, key=lambda x: -x, reverse=True)) l.sort(key=lambda x: -x, reverse=True) print(l) print(l == sorted(l, key=lambda x: -x, reverse=True)) print(sorted(l, reverse=True)) l.sort(reverse=True) print(l) print(l == sorted(l, reverse=True)) print(sorted(l, reverse=False)) l.sort(reverse=False) print(l) print(l == sorted(l, reverse=False))
Add tests for sorted() function and check that sorted(list) produces same output as list.sort()
Add tests for sorted() function and check that sorted(list) produces same output as list.sort()
Python
mit
MrSurly/micropython-esp32,KISSMonX/micropython,mpalomer/micropython,mianos/micropython,ernesto-g/micropython,dxxb/micropython,tuc-osg/micropython,lbattraw/micropython,turbinenreiter/micropython,swegener/micropython,swegener/micropython,praemdonck/micropython,SungEun-Steve-Kim/test-mp,pfalcon/micropython,pozetroninc/micropython,mgyenik/micropython,vriera/micropython,noahwilliamsson/micropython,aethaniel/micropython,orionrobots/micropython,rubencabrera/micropython,vitiral/micropython,torwag/micropython,omtinez/micropython,drrk/micropython,lbattraw/micropython,cnoviello/micropython,galenhz/micropython,TDAbboud/micropython,jimkmc/micropython,ganshun666/micropython,mhoffma/micropython,adafruit/circuitpython,drrk/micropython,oopy/micropython,rubencabrera/micropython,tralamazza/micropython,MrSurly/micropython-esp32,alex-robbins/micropython,cwyark/micropython,slzatz/micropython,turbinenreiter/micropython,pozetroninc/micropython,HenrikSolver/micropython,ganshun666/micropython,pramasoul/micropython,aitjcize/micropython,noahwilliamsson/micropython,matthewelse/micropython,hosaka/micropython,martinribelotta/micropython,PappaPeppar/micropython,redbear/micropython,slzatz/micropython,AriZuu/micropython,alex-robbins/micropython,MrSurly/micropython-esp32,xuxiaoxin/micropython,misterdanb/micropython,vitiral/micropython,lbattraw/micropython,emfcamp/micropython,mianos/micropython,skybird6672/micropython,turbinenreiter/micropython,hiway/micropython,ryannathans/micropython,trezor/micropython,heisewangluo/micropython,xhat/micropython,ChuckM/micropython,tuc-osg/micropython,danicampora/micropython,methoxid/micropystat,toolmacher/micropython,adafruit/circuitpython,pramasoul/micropython,tdautc19841202/micropython,martinribelotta/micropython,Peetz0r/micropython-esp32,warner83/micropython,adafruit/micropython,tobbad/micropython,TDAbboud/micropython,adamkh/micropython,Peetz0r/micropython-esp32,pfalcon/micropython,ruffy91/micropython,alex-march/micropython,adafruit/micropython,orionrobots/micropython,mgyenik/micropython,martinribelotta/micropython,tdautc19841202/micropython,mianos/micropython,henriknelson/micropython,orionrobots/micropython,tdautc19841202/micropython,ryannathans/micropython,neilh10/micropython,lowRISC/micropython,tobbad/micropython,lbattraw/micropython,ernesto-g/micropython,ChuckM/micropython,adamkh/micropython,dxxb/micropython,tdautc19841202/micropython,cloudformdesign/micropython,ernesto-g/micropython,alex-march/micropython,HenrikSolver/micropython,redbear/micropython,mpalomer/micropython,xuxiaoxin/micropython,SHA2017-badge/micropython-esp32,adamkh/micropython,jlillest/micropython,suda/micropython,trezor/micropython,TDAbboud/micropython,Vogtinator/micropython,jimkmc/micropython,chrisdearman/micropython,lbattraw/micropython,aethaniel/micropython,redbear/micropython,pozetroninc/micropython,jmarcelino/pycom-micropython,stonegithubs/micropython,xuxiaoxin/micropython,Vogtinator/micropython,utopiaprince/micropython,turbinenreiter/micropython,blazewicz/micropython,orionrobots/micropython,dxxb/micropython,skybird6672/micropython,skybird6672/micropython,ceramos/micropython,supergis/micropython,xyb/micropython,paul-xxx/micropython,tdautc19841202/micropython,adafruit/micropython,selste/micropython,neilh10/micropython,danicampora/micropython,AriZuu/micropython,utopiaprince/micropython,misterdanb/micropython,oopy/micropython,galenhz/micropython,supergis/micropython,aethaniel/micropython,noahwilliamsson/micropython,mpalomer/micropython,pozetroninc/micropython,dmazzella/micropython,neilh10/micropython,ruffy91/micropython,ernesto-g/micropython,cnoviello/micropython,vriera/micropython,PappaPeppar/micropython,matthewelse/micropython,emfcamp/micropython,neilh10/micropython,mgyenik/micropython,firstval/micropython,jimkmc/micropython,utopiaprince/micropython,hiway/micropython,pramasoul/micropython,skybird6672/micropython,stonegithubs/micropython,tralamazza/micropython,ryannathans/micropython,blmorris/micropython,adamkh/micropython,tralamazza/micropython,dinau/micropython,jmarcelino/pycom-micropython,ceramos/micropython,torwag/micropython,dmazzella/micropython,ericsnowcurrently/micropython,infinnovation/micropython,ahotam/micropython,adamkh/micropython,supergis/micropython,noahwilliamsson/micropython,tobbad/micropython,orionrobots/micropython,SungEun-Steve-Kim/test-mp,paul-xxx/micropython,paul-xxx/micropython,ruffy91/micropython,emfcamp/micropython,dmazzella/micropython,kostyll/micropython,mpalomer/micropython,oopy/micropython,selste/micropython,hiway/micropython,trezor/micropython,pfalcon/micropython,deshipu/micropython,SHA2017-badge/micropython-esp32,alex-robbins/micropython,dinau/micropython,omtinez/micropython,tuc-osg/micropython,warner83/micropython,swegener/micropython,chrisdearman/micropython,mgyenik/micropython,rubencabrera/micropython,martinribelotta/micropython,toolmacher/micropython,mpalomer/micropython,hosaka/micropython,Timmenem/micropython,oopy/micropython,trezor/micropython,pramasoul/micropython,matthewelse/micropython,tuc-osg/micropython,dhylands/micropython,henriknelson/micropython,tuc-osg/micropython,alex-robbins/micropython,omtinez/micropython,KISSMonX/micropython,cnoviello/micropython,paul-xxx/micropython,drrk/micropython,toolmacher/micropython,hosaka/micropython,vitiral/micropython,vriera/micropython,ruffy91/micropython,dxxb/micropython,tralamazza/micropython,micropython/micropython-esp32,MrSurly/micropython-esp32,hiway/micropython,Timmenem/micropython,ganshun666/micropython,feilongfl/micropython,stonegithubs/micropython,xyb/micropython,adafruit/circuitpython,dinau/micropython,turbinenreiter/micropython,noahchense/micropython,infinnovation/micropython,EcmaXp/micropython,supergis/micropython,kostyll/micropython,jimkmc/micropython,supergis/micropython,aitjcize/micropython,vitiral/micropython,ericsnowcurrently/micropython,galenhz/micropython,hosaka/micropython,jimkmc/micropython,cwyark/micropython,suda/micropython,praemdonck/micropython,lowRISC/micropython,xyb/micropython,micropython/micropython-esp32,omtinez/micropython,heisewangluo/micropython,HenrikSolver/micropython,aethaniel/micropython,ahotam/micropython,puuu/micropython,suda/micropython,ernesto-g/micropython,misterdanb/micropython,cnoviello/micropython,cloudformdesign/micropython,noahchense/micropython,HenrikSolver/micropython,firstval/micropython,chrisdearman/micropython,blazewicz/micropython,cnoviello/micropython,suda/micropython,AriZuu/micropython,SungEun-Steve-Kim/test-mp,AriZuu/micropython,puuu/micropython,deshipu/micropython,xhat/micropython,HenrikSolver/micropython,kerneltask/micropython,pramasoul/micropython,galenhz/micropython,MrSurly/micropython,stonegithubs/micropython,jmarcelino/pycom-micropython,SHA2017-badge/micropython-esp32,ChuckM/micropython,Vogtinator/micropython,warner83/micropython,KISSMonX/micropython,mianos/micropython,xyb/micropython,firstval/micropython,martinribelotta/micropython,cwyark/micropython,mhoffma/micropython,MrSurly/micropython,jlillest/micropython,vitiral/micropython,deshipu/micropython,pozetroninc/micropython,heisewangluo/micropython,firstval/micropython,pfalcon/micropython,lowRISC/micropython,TDAbboud/micropython,KISSMonX/micropython,jlillest/micropython,EcmaXp/micropython,adafruit/circuitpython,kerneltask/micropython,puuu/micropython,omtinez/micropython,praemdonck/micropython,Vogtinator/micropython,puuu/micropython,warner83/micropython,aethaniel/micropython,oopy/micropython,mhoffma/micropython,methoxid/micropystat,puuu/micropython,redbear/micropython,rubencabrera/micropython,hiway/micropython,dhylands/micropython,bvernoux/micropython,bvernoux/micropython,emfcamp/micropython,feilongfl/micropython,EcmaXp/micropython,xhat/micropython,PappaPeppar/micropython,SHA2017-badge/micropython-esp32,noahchense/micropython,noahchense/micropython,swegener/micropython,mhoffma/micropython,cwyark/micropython,ryannathans/micropython,aitjcize/micropython,jmarcelino/pycom-micropython,methoxid/micropystat,xuxiaoxin/micropython,ceramos/micropython,kostyll/micropython,neilh10/micropython,ericsnowcurrently/micropython,infinnovation/micropython,adafruit/circuitpython,dhylands/micropython,mianos/micropython,EcmaXp/micropython,xyb/micropython,deshipu/micropython,trezor/micropython,TDAbboud/micropython,praemdonck/micropython,mhoffma/micropython,ChuckM/micropython,SHA2017-badge/micropython-esp32,aitjcize/micropython,bvernoux/micropython,kerneltask/micropython,KISSMonX/micropython,selste/micropython,utopiaprince/micropython,bvernoux/micropython,alex-march/micropython,alex-robbins/micropython,kostyll/micropython,blmorris/micropython,ganshun666/micropython,paul-xxx/micropython,kerneltask/micropython,hosaka/micropython,kerneltask/micropython,ryannathans/micropython,ceramos/micropython,vriera/micropython,adafruit/circuitpython,slzatz/micropython,xhat/micropython,swegener/micropython,toolmacher/micropython,torwag/micropython,xhat/micropython,danicampora/micropython,suda/micropython,lowRISC/micropython,heisewangluo/micropython,rubencabrera/micropython,torwag/micropython,SungEun-Steve-Kim/test-mp,cwyark/micropython,lowRISC/micropython,dinau/micropython,micropython/micropython-esp32,dhylands/micropython,PappaPeppar/micropython,danicampora/micropython,noahwilliamsson/micropython,Peetz0r/micropython-esp32,infinnovation/micropython,ganshun666/micropython,jlillest/micropython,ceramos/micropython,henriknelson/micropython,firstval/micropython,cloudformdesign/micropython,MrSurly/micropython-esp32,drrk/micropython,Timmenem/micropython,blazewicz/micropython,selste/micropython,matthewelse/micropython,Peetz0r/micropython-esp32,torwag/micropython,galenhz/micropython,alex-march/micropython,mgyenik/micropython,skybird6672/micropython,noahchense/micropython,misterdanb/micropython,feilongfl/micropython,heisewangluo/micropython,misterdanb/micropython,cloudformdesign/micropython,ahotam/micropython,ahotam/micropython,Timmenem/micropython,vriera/micropython,cloudformdesign/micropython,emfcamp/micropython,matthewelse/micropython,blmorris/micropython,slzatz/micropython,dxxb/micropython,feilongfl/micropython,feilongfl/micropython,MrSurly/micropython,blazewicz/micropython,toolmacher/micropython,Timmenem/micropython,infinnovation/micropython,stonegithubs/micropython,dhylands/micropython,kostyll/micropython,Peetz0r/micropython-esp32,blmorris/micropython,ericsnowcurrently/micropython,praemdonck/micropython,blazewicz/micropython,jmarcelino/pycom-micropython,ahotam/micropython,ruffy91/micropython,drrk/micropython,chrisdearman/micropython,Vogtinator/micropython,alex-march/micropython,methoxid/micropystat,pfalcon/micropython,jlillest/micropython,micropython/micropython-esp32,MrSurly/micropython,adafruit/micropython,warner83/micropython,EcmaXp/micropython,MrSurly/micropython,slzatz/micropython,AriZuu/micropython,blmorris/micropython,bvernoux/micropython,danicampora/micropython,henriknelson/micropython,SungEun-Steve-Kim/test-mp,ericsnowcurrently/micropython,matthewelse/micropython,ChuckM/micropython,PappaPeppar/micropython,adafruit/micropython,dinau/micropython,selste/micropython,henriknelson/micropython,tobbad/micropython,xuxiaoxin/micropython,dmazzella/micropython,utopiaprince/micropython,chrisdearman/micropython,redbear/micropython,deshipu/micropython,methoxid/micropystat,micropython/micropython-esp32,tobbad/micropython
--- +++ @@ -1,13 +1,28 @@ l = [1, 3, 2, 5] + print(l) +print(sorted(l)) l.sort() print(l) +print(l == sorted(l)) + +print(sorted(l, key=lambda x: -x)) l.sort(key=lambda x: -x) print(l) +print(l == sorted(l, key=lambda x: -x)) + +print(sorted(l, key=lambda x: -x, reverse=True)) l.sort(key=lambda x: -x, reverse=True) print(l) +print(l == sorted(l, key=lambda x: -x, reverse=True)) + +print(sorted(l, reverse=True)) l.sort(reverse=True) print(l) +print(l == sorted(l, reverse=True)) + +print(sorted(l, reverse=False)) l.sort(reverse=False) print(l) +print(l == sorted(l, reverse=False))
2d688f97b9869fdfed9237b91fdce287278e3c6c
wsgi.py
wsgi.py
import os from elasticsearch_raven.transport import ElasticsearchTransport from elasticsearch_raven.utils import get_index host = os.environ.get('ELASTICSEARCH_HOST', 'localhost:9200') transport = ElasticsearchTransport(host) def application(environ, start_response): index = get_index(environ) transport.send(environ['wsgi.input'].read(), index) status = '200 OK' response_headers = [('Content-Type', 'text/plain')] start_response(status, response_headers) return [''.encode('utf-8')]
import os from queue import Queue from threading import Thread from elasticsearch_raven.transport import ElasticsearchTransport from elasticsearch_raven.utils import get_index host = os.environ.get('ELASTICSEARCH_HOST', 'localhost:9200') transport = ElasticsearchTransport(host) blocking_queue = Queue() def send(): while True: data, index = blocking_queue.get() transport.send(data, index) blocking_queue.task_done() sender = Thread(target=send) sender.daemon = True sender.start() def application(environ, start_response): index = get_index(environ) length = int(environ.get('CONTENT_LENGTH', '0')) data = environ['wsgi.input'].read(length) blocking_queue.put((data, index)) status = '200 OK' response_headers = [('Content-Type', 'text/plain')] start_response(status, response_headers) return [''.encode('utf-8')]
Send data to elasticsearch asynchronously.
Send data to elasticsearch asynchronously.
Python
mit
socialwifi/elasticsearch-raven,pozytywnie/elasticsearch-raven,serathius/elasticsearch-raven
--- +++ @@ -1,15 +1,32 @@ import os +from queue import Queue +from threading import Thread from elasticsearch_raven.transport import ElasticsearchTransport from elasticsearch_raven.utils import get_index host = os.environ.get('ELASTICSEARCH_HOST', 'localhost:9200') transport = ElasticsearchTransport(host) +blocking_queue = Queue() + + +def send(): + while True: + data, index = blocking_queue.get() + transport.send(data, index) + blocking_queue.task_done() + + +sender = Thread(target=send) +sender.daemon = True +sender.start() def application(environ, start_response): index = get_index(environ) - transport.send(environ['wsgi.input'].read(), index) + length = int(environ.get('CONTENT_LENGTH', '0')) + data = environ['wsgi.input'].read(length) + blocking_queue.put((data, index)) status = '200 OK' response_headers = [('Content-Type', 'text/plain')]
d703d7cb8d75a5c660beabccdd0082794a8471d1
edisgo/tools/networkx_helper.py
edisgo/tools/networkx_helper.py
from networkx import OrderedGraph def translate_df_to_graph(buses_df, lines_df, transformers_df=None): graph = OrderedGraph() buses = buses_df.index # add nodes graph.add_nodes_from(buses) # add branches branches = [] for line_name, line in lines_df.iterrows(): branches.append( ( line.bus0, line.bus1, {"branch_name": line_name, "length": line.length}, ) ) if transformers_df is not None: for trafo_name, trafo in transformers_df.iterrows(): branches.append( ( trafo.bus0, trafo.bus1, {"branch_name": trafo_name, "length": 0}, ) ) graph.add_edges_from(branches) return graph
from networkx import Graph def translate_df_to_graph(buses_df, lines_df, transformers_df=None): graph = Graph() buses = [] for bus_name, bus in buses_df.iterrows(): pos = (bus.x, bus.y) buses.append((bus_name, {'pos': pos})) # add nodes graph.add_nodes_from(buses) # add branches branches = [] for line_name, line in lines_df.iterrows(): branches.append( ( line.bus0, line.bus1, {"branch_name": line_name, "length": line.length}, ) ) if transformers_df is not None: for trafo_name, trafo in transformers_df.iterrows(): branches.append( ( trafo.bus0, trafo.bus1, {"branch_name": trafo_name, "length": 0}, ) ) graph.add_edges_from(branches) return graph
Include the position into the graph
Include the position into the graph
Python
agpl-3.0
openego/eDisGo,openego/eDisGo
--- +++ @@ -1,11 +1,14 @@ -from networkx import OrderedGraph +from networkx import Graph def translate_df_to_graph(buses_df, lines_df, transformers_df=None): - graph = OrderedGraph() + graph = Graph() - buses = buses_df.index - # add nodes + buses = [] + for bus_name, bus in buses_df.iterrows(): + pos = (bus.x, bus.y) + buses.append((bus_name, {'pos': pos})) + # add nodes graph.add_nodes_from(buses) # add branches branches = []
78da548cbf7646ea1d67859cbd0946d6931b2b0d
ad-hoc-scripts/lift.py
ad-hoc-scripts/lift.py
#! /usr/bin/env python3 import sys import json for arg in sys.argv[1:]: with open(arg) as f: equajson = json.load(f) field_list = [ "always-an-integer", "always-positive", "always-dimensionless", "bound-variable", "fixed-constant", "special-function" ] try: for term in equajson["unicode-pretty-print"]["terms"]: if any(x in field_list for x in term.keys()): term["classification"] = {} del_list = [] for field in term.keys(): if field in field_list: term["classification"][field] = term[field] del_list.append(field) for field in del_list: del term[field] except KeyError: pass with open(sys.argv[1], 'w') as f: json.dump(equajson, f, indent=4, separators=(',', ': '), ensure_ascii=False, sort_keys=True)
#! /usr/bin/env python3 import sys import json for arg in sys.argv[1:]: with open(arg) as f: equajson = json.load(f) field_list = [ "always-an-integer", "always-positive", "always-dimensionless", "bound-variable", "fixed-constant", "special-function" ] try: for term in equajson["unicode-pretty-print"]["terms"]: if any(x in field_list for x in term.keys()): term["classification"] = {} del_list = [] for field in term.keys(): if field in field_list: term["classification"][field] = term[field] del_list.append(field) for field in del_list: del term[field] except KeyError: pass with open(arg, 'w') as f: json.dump(equajson, f, indent=4, separators=(',', ': '), ensure_ascii=False, sort_keys=True)
Fix serious file overwriting problem.
Fix serious file overwriting problem.
Python
mit
nbeaver/equajson
--- +++ @@ -29,5 +29,5 @@ except KeyError: pass - with open(sys.argv[1], 'w') as f: + with open(arg, 'w') as f: json.dump(equajson, f, indent=4, separators=(',', ': '), ensure_ascii=False, sort_keys=True)
80271752183c3f71fba9139bb77466427bd48a0a
alfred_listener/__main__.py
alfred_listener/__main__.py
#!/usr/bin/env python import os from argh import arg, ArghParser from functools import wraps CONFIG = os.environ.get('ALFRED_LISTENER_CONFIG') def with_app(func, args): if CONFIG is None: raise RuntimeError('ALFRED_LISTENER_CONFIG env variable is not set.') @wraps(func) def wrapper(*args, **kwargs): from alfred_listener import create_app app = create_app(config) return func(app, *args, **kwargs) return wrapper @arg('--host', default='127.0.0.1', help='the host') @arg('--port', default=5000, help='the port') @with_app def runserver(app, args): app.run(args.host, args.port) @with_app def shell(app, args): from alfred_listener.helpers import get_shell with app.test_request_context(): sh = get_shell() sh(app=app) def main(): parser = ArghParser() parser.add_commands([runserver, shell]) parser.dispatch() if __name__ == '__main__': main()
#!/usr/bin/env python import os from argh import arg, ArghParser from functools import wraps def with_app(func): @arg('--config', help='Path to config file') @wraps(func) def wrapper(*args, **kwargs): config = args[0].config from alfred_listener import create_app app = create_app(config) return func(app, *args, **kwargs) return wrapper @arg('--host', default='127.0.0.1', help='the host') @arg('--port', default=5000, help='the port') @with_app def runserver(app, args): app.run(args.host, args.port) @with_app def shell(app, args): from alfred_listener.helpers import get_shell with app.test_request_context(): sh = get_shell() sh(app=app) def main(): parser = ArghParser() parser.add_commands([runserver, shell]) parser.dispatch() if __name__ == '__main__': main()
Change the way to determine path to config file
Change the way to determine path to config file
Python
isc
alfredhq/alfred-listener
--- +++ @@ -5,14 +5,11 @@ from functools import wraps -CONFIG = os.environ.get('ALFRED_LISTENER_CONFIG') - - -def with_app(func, args): - if CONFIG is None: - raise RuntimeError('ALFRED_LISTENER_CONFIG env variable is not set.') +def with_app(func): + @arg('--config', help='Path to config file') @wraps(func) def wrapper(*args, **kwargs): + config = args[0].config from alfred_listener import create_app app = create_app(config) return func(app, *args, **kwargs)
30b64d934d12510e2021aa3bf2e80f5837b88f35
mnist_sequence_cli.py
mnist_sequence_cli.py
from __future__ import print_function from mnist_sequence_api import MNIST_Sequence_API import sys def main(): arguments = list(sys.argv[1:]) if len(arguments) == 4: sequence, min_spacing = list(map(int, list(arguments[0])), int(arguments[1])) max_spacing, image_width = int(arguments[2]), int(arguments[3]) api_object = MNIST_Sequence_API() img_data = api_object.generate_mnist_sequence(sequence, (min_spacing, max_spacing), image_width) api_object.save_image(img_data, sequence) else: print("Incorrect number of arguments.") print("Usage: python mnist_sequence_cli.py <sequence(no spaces) " + "min_spacing max_spacing image_width>") if __name__ == "__main__": main()
from __future__ import print_function from mnist_sequence_api import MNIST_Sequence_API import sys def main(): arguments = list(sys.argv[1:]) if len(arguments) == 4: sequence, min_spacing = list(map(int, list(arguments[0]))), int(arguments[1]) max_spacing, image_width = int(arguments[2]), int(arguments[3]) api_object = MNIST_Sequence_API() img_data = api_object.generate_mnist_sequence(sequence, (min_spacing, max_spacing), image_width) api_object.save_image(img_data, sequence) else: print("Incorrect number of arguments.") print("Usage: python mnist_sequence_cli.py <sequence(no spaces) " + "min_spacing max_spacing image_width>") if __name__ == "__main__": main()
Fix argument inputs to list()
Fix argument inputs to list()
Python
mit
ankitaggarwal011/MNIST-Sequence
--- +++ @@ -6,7 +6,7 @@ def main(): arguments = list(sys.argv[1:]) if len(arguments) == 4: - sequence, min_spacing = list(map(int, list(arguments[0])), int(arguments[1])) + sequence, min_spacing = list(map(int, list(arguments[0]))), int(arguments[1]) max_spacing, image_width = int(arguments[2]), int(arguments[3]) api_object = MNIST_Sequence_API() img_data = api_object.generate_mnist_sequence(sequence, (min_spacing, max_spacing), image_width)
8fdd73b7b8b907da264eac26ea52eef2bb8f6ce7
ml/img/load_images.py
ml/img/load_images.py
# ============================================================================== # LOAD_IMAGE_AS_ARRAY # ============================================================================== import scipy.misc def load_image_as_array(f, rescale=None): """ Given a filepath to an image file, it loads an image as a numpy array. Optionally resize the images to [width, height]""" img = scipy.misc.imread(f) if rescale: width, height = rescale img = scipy.misc.imresize(img, (height,width)) return img
# ============================================================================== # LOAD_IMAGE_AS_ARRAY # ============================================================================== # USING SCIPY (with i think PIL on the backend) import scipy.misc def load_image_as_array(f, rescale=None): """ Given a filepath to an image file, it loads an image as a numpy array. Optionally resize the images to [width, height]""" img = scipy.misc.imread(f) if rescale: width, height = rescale img = scipy.misc.imresize(img, (height,width)) return img # USING OPEN CV def load_image_as_array(f, rescale=None): # TODO: Check the order of the dimensions for resizing in open cv img = cv2.imread(f, cv2.IMREAD_COLOR) img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB) # Convert to RGB if rescale: img = cv2.resize(img, rescale) return img
Add opencv version of load_image_as_array()
FEAT: Add opencv version of load_image_as_array()
Python
apache-2.0
ronrest/convenience_py,ronrest/convenience_py
--- +++ @@ -1,6 +1,7 @@ # ============================================================================== # LOAD_IMAGE_AS_ARRAY # ============================================================================== +# USING SCIPY (with i think PIL on the backend) import scipy.misc def load_image_as_array(f, rescale=None): """ Given a filepath to an image file, it loads an image as a numpy array. @@ -10,3 +11,15 @@ width, height = rescale img = scipy.misc.imresize(img, (height,width)) return img + + +# USING OPEN CV +def load_image_as_array(f, rescale=None): + # TODO: Check the order of the dimensions for resizing in open cv + img = cv2.imread(f, cv2.IMREAD_COLOR) + img = cv2.cvtColor(img, cv2.COLOR_BGR2RGB) # Convert to RGB + if rescale: + img = cv2.resize(img, rescale) + return img + +
9516caa0c62289c89b605b9b8a34622a0bb54e2b
tests/70_program_libpfasst_swe_sphere_timestepper_convergence_l1/postprocessing_pickle.py
tests/70_program_libpfasst_swe_sphere_timestepper_convergence_l1/postprocessing_pickle.py
#! /usr/bin/env python3 import sys import math import glob from mule_local.postprocessing.pickle_SphereDataPhysicalDiff import * from mule.exec_program import * pickle_SphereDataPhysicalDiff()
#! /usr/bin/env python3 import sys import math import glob from mule_local.postprocessing.pickle_SphereDataSpectralDiff import * from mule.exec_program import * pickle_SphereDataSpectralDiff()
Use SphereDataSpectral instead of SphereDataPhysical
Use SphereDataSpectral instead of SphereDataPhysical The script for physical data requires CSV files as default output files, whereas the spectral script uses .sweet binary files as default. Since libpfasst_swe_sphere's CSV output is not in the same format as swe_sphere's CSV output, the CSV parsing does not work --> this is the easiest way to use the (working) binary output files.
Python
mit
schreiberx/sweet,schreiberx/sweet,schreiberx/sweet,schreiberx/sweet
--- +++ @@ -4,7 +4,7 @@ import math import glob -from mule_local.postprocessing.pickle_SphereDataPhysicalDiff import * +from mule_local.postprocessing.pickle_SphereDataSpectralDiff import * from mule.exec_program import * -pickle_SphereDataPhysicalDiff() +pickle_SphereDataSpectralDiff()
33f4036825c6ff4d9df0038471727648e0df100d
feder/virus_scan/engine/base.py
feder/virus_scan/engine/base.py
from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), self.signer.sign(self.name), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
import urllib.parse from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner class BaseEngine: def __init__(self): self.signer = TimestampSigner() def get_webhook_url(self): return "{}://{}{}?token={}".format( "https", get_current_site(None).domain, reverse("virus_scan:webhook"), urllib.parse.quote(self.signer.sign(self.name)), ) def send_scan(self, this_file, filename): raise NotImplementedError( "Provide 'send' in {name}".format(name=self.__class__.__name__) ) def receive_scan(self, engine_id): raise NotImplementedError( "Provide 'receive_scan' in {name}".format(name=self.__class__.__name__) )
Fix urlencode in webhook url
Fix urlencode in webhook url
Python
mit
watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder,watchdogpolska/feder
--- +++ @@ -1,3 +1,5 @@ +import urllib.parse + from django.contrib.sites.shortcuts import get_current_site from django.urls import reverse from django.core.signing import TimestampSigner @@ -12,7 +14,7 @@ "https", get_current_site(None).domain, reverse("virus_scan:webhook"), - self.signer.sign(self.name), + urllib.parse.quote(self.signer.sign(self.name)), ) def send_scan(self, this_file, filename):
d618f430c143874011b70afe0a4fa62c06f5e28c
md5bot.py
md5bot.py
""" md5bot.py -- Twitter bot that tweets the current time as an md5 value """ import time import hashlib import tweepy CONSUMER_KEY = 'xxxxxxxxxxxx' CONSUMER_SECRET = 'xxxxxxxxxxxx' ACCESS_KEY = 'xxxxxxxxxxxx' ACCESS_SECRET = 'xxxxxxxxxxxx' auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET) auth.set_access_token(ACCESS_KEY, ACCESS_SECRET) api = tweepy.API(auth) def checkTime(): ts = time.time() m = hashlib.md5() m.update(str(ts)) return m.hexdigest() def postStatus(): currentTime = checkTime() api.update_status(currentTime) int = 0 while int == 0: postStatus() time.sleep(3600) #Sleep for one hour
#!/usr/bin/env python __author__ = "Patrick Guelcher" __copyright__ = "(C) 2016 Patrick Guelcher" __license__ = "MIT" __version__ = "1.0" """ A bot for Twitter that checks the time and then posts it as an md5 hash value. """ import time import hashlib import tweepy # Configuration (Twitter API Settings) CONSUMER_KEY = 'npfl47weJ6vSn3MRXUq342dMB' CONSUMER_SECRET = 'dKv6zrr7ExIWAtVE3gWG4xZFs3LziZaeHvmycTHkttGC3etP4d' ACCESS_TOKEN = '2489159084-t6A6zXVZSJFdZYP8jb78Mat8Lg3TfnIdffBgUTs' ACCESS_SECRET = '0C83TOgZ4WE00zWuDxVT2TS6E5sVo0Bp0P1IpRn2ipNhD' sleep_time = 3600 # Time is in seconds (Default 3600) # Do not edit beyond this comment def main(): index = 0 while index == 0: post_status() time.sleep(sleep_time) # Sleep for one hour def check_time(): time_stamp = str(time.time()) encode_time = time_stamp.encode('utf-8') md = hashlib.md5() md.update(encode_time) return md.hexdigest() def post_status(): auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET) auth.set_access_token(ACCESS_TOKEN, ACCESS_SECRET) api = tweepy.API(auth) current_time = check_time() api.update_status(current_time) print(current_time) if __name__ == '__main__': main()
Clean code and conform to python conventions
Clean code and conform to python conventions Some things are still a bit weird, mostly due to my limited knowledge of Python. Also fixed code to conform to Python naming conventions for variables/functions.
Python
mit
aerovolts/python-scripts
--- +++ @@ -1,30 +1,47 @@ +#!/usr/bin/env python + +__author__ = "Patrick Guelcher" +__copyright__ = "(C) 2016 Patrick Guelcher" +__license__ = "MIT" +__version__ = "1.0" + """ -md5bot.py -- Twitter bot that tweets the current time as an md5 value +A bot for Twitter that checks the time and then posts it as an md5 hash value. """ import time import hashlib import tweepy -CONSUMER_KEY = 'xxxxxxxxxxxx' -CONSUMER_SECRET = 'xxxxxxxxxxxx' -ACCESS_KEY = 'xxxxxxxxxxxx' -ACCESS_SECRET = 'xxxxxxxxxxxx' -auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET) -auth.set_access_token(ACCESS_KEY, ACCESS_SECRET) -api = tweepy.API(auth) +# Configuration (Twitter API Settings) +CONSUMER_KEY = 'npfl47weJ6vSn3MRXUq342dMB' +CONSUMER_SECRET = 'dKv6zrr7ExIWAtVE3gWG4xZFs3LziZaeHvmycTHkttGC3etP4d' +ACCESS_TOKEN = '2489159084-t6A6zXVZSJFdZYP8jb78Mat8Lg3TfnIdffBgUTs' +ACCESS_SECRET = '0C83TOgZ4WE00zWuDxVT2TS6E5sVo0Bp0P1IpRn2ipNhD' +sleep_time = 3600 # Time is in seconds (Default 3600) -def checkTime(): - ts = time.time() - m = hashlib.md5() - m.update(str(ts)) - return m.hexdigest() +# Do not edit beyond this comment +def main(): + index = 0 + while index == 0: + post_status() + time.sleep(sleep_time) # Sleep for one hour -def postStatus(): - currentTime = checkTime() - api.update_status(currentTime) +def check_time(): + time_stamp = str(time.time()) + encode_time = time_stamp.encode('utf-8') + md = hashlib.md5() + md.update(encode_time) + return md.hexdigest() -int = 0 -while int == 0: - postStatus() - time.sleep(3600) #Sleep for one hour +def post_status(): + auth = tweepy.OAuthHandler(CONSUMER_KEY, CONSUMER_SECRET) + auth.set_access_token(ACCESS_TOKEN, ACCESS_SECRET) + api = tweepy.API(auth) + + current_time = check_time() + api.update_status(current_time) + print(current_time) + +if __name__ == '__main__': + main()
1a39eea8225ebdf7f654df9ba5b87479e9dbc867
minify.py
minify.py
""" Copyright 2016 Dee Reddy """ import sys import re args = sys.argv[1:] def minify(filepath, comments=False): """ Minifies/uglifies file :param file_: comments: Boolean. If False, deletes comments during output. :return: Minified string. """ pattern = re.compile(r""" \s | # matches all whitespace characters OR ( # /\* # /* [ \w\s (?=\*) :@!"'~\.\^\$\+\?\{\}\[\]\\\|\(\) ]* # AND \*/ # */ ) # | //.*\n # OR any character from // until end-line (inclusive) """, re.VERBOSE) with open(filepath, "r") as file_: temp = [] for line in file_: temp.append(line) output = ''.join(temp) return pattern.sub('', output) if __name__ == "__main__": print(minify('./test/stylesheet.css'))
""" Copyright 2016 Dee Reddy """ import sys import re args = sys.argv[1:] def minify(input_path, output_path, comments=False): """ Minifies/uglifies file args: input_path: input file path output_path: write-out file path comments: Boolean. If False, deletes comments during output. returns: Minified string. example: `$ python minify.py ./src/styles.css ./src/output.css` """ pattern = re.compile(r""" \s | # matches all whitespace characters OR ( # /\* # /* AND [ # 0 or more of any character \w\s # (?=\*) # (positive lookahead: doesn't make * part of the match) :@!"'~,#%&-=;<>` # \.\^\$\+\{\[\]\\\| # ]* # \*/ # AND */ ) # | //.*\n # OR any character from // until end-line (inclusive) """, re.VERBOSE) # read file and apply regex: with open(input_path, "r") as file_in: temp = [] for line in file_in: temp.append(line) output = ''.join(temp) output = pattern.sub('', output) # write to file: # (`w+` mode: writing/reading; overwrites existing files; creates file if doesn't exit) with open(output_path, "w+") as file_out: file_out.write(output) ############################# # Main # ############################# if __name__ == "__main__": # specify input and output paths in args: minify(args[0], args[1])
Update regex. Add file write-out.
feat: Update regex. Add file write-out. Regex is bugged, however. Need to fix regex pattern.
Python
apache-2.0
Deesus/Punt
--- +++ @@ -8,36 +8,49 @@ args = sys.argv[1:] -def minify(filepath, comments=False): +def minify(input_path, output_path, comments=False): """ Minifies/uglifies file - :param - file_: + args: + input_path: input file path + output_path: write-out file path comments: Boolean. If False, deletes comments during output. - :return: + returns: Minified string. + example: + `$ python minify.py ./src/styles.css ./src/output.css` """ pattern = re.compile(r""" - \s | # matches all whitespace characters OR - ( # - /\* # /* - [ - \w\s - (?=\*) - :@!"'~\.\^\$\+\?\{\}\[\]\\\|\(\) - ]* # AND - \*/ # */ - ) # - | //.*\n # OR any character from // until end-line (inclusive) + \s | # matches all whitespace characters OR + ( # + /\* # /* AND + [ # 0 or more of any character + \w\s # + (?=\*) # (positive lookahead: doesn't make * part of the match) + :@!"'~,#%&-=;<>` # + \.\^\$\+\{\[\]\\\| # + ]* # + \*/ # AND */ + ) # + | //.*\n # OR any character from // until end-line (inclusive) """, re.VERBOSE) - with open(filepath, "r") as file_: + # read file and apply regex: + with open(input_path, "r") as file_in: temp = [] - for line in file_: + for line in file_in: temp.append(line) output = ''.join(temp) - return pattern.sub('', output) + output = pattern.sub('', output) + # write to file: + # (`w+` mode: writing/reading; overwrites existing files; creates file if doesn't exit) + with open(output_path, "w+") as file_out: + file_out.write(output) +############################# +# Main # +############################# if __name__ == "__main__": - print(minify('./test/stylesheet.css')) + # specify input and output paths in args: + minify(args[0], args[1])
f7a57135ec328063eb4d2d5c4a035994a85ad290
modules/piperename.py
modules/piperename.py
# piperename.py # from pipe2py import util def pipe_rename(context, _INPUT, conf, **kwargs): """This operator renames or copies fields in the input source. Keyword arguments: context -- pipeline context _INPUT -- source generator kwargs -- other inputs, e.g. to feed terminals for rule values conf: RULE -- rules - each rule comprising (op, field, newval) Yields (_OUTPUT): source items after copying/renaming """ rules = [] for rule in conf['RULE']: newval = util.get_value(rule['newval'], kwargs) #todo use subkey? rules.append((rule['op']['value'], rule['field']['value'], newval)) for item in _INPUT: for rule in rules: item[rule[2]] = item[rule[1]] if rule[0] == 'rename': del item[rule[1]] yield item
# piperename.py # from pipe2py import util def pipe_rename(context, _INPUT, conf, **kwargs): """This operator renames or copies fields in the input source. Keyword arguments: context -- pipeline context _INPUT -- source generator kwargs -- other inputs, e.g. to feed terminals for rule values conf: RULE -- rules - each rule comprising (op, field, newval) Yields (_OUTPUT): source items after copying/renaming """ rules = [] for rule in conf['RULE']: newval = util.get_value(rule['newval'], kwargs) #todo use subkey? rules.append((rule['op']['value'], rule['field']['value'], newval)) for item in _INPUT: for rule in rules: #Map names with dot notation onto nested dictionaries, e.g. 'a.content' -> ['a']['content'] #todo: optimise by pre-calculating splits # and if this logic is stable, wrap in util functions and use everywhere items are accessed reduce(lambda i,k:i.get(k), [item] + rule[2].split('.')[:-1])[rule[2].split('.')[-1]] = reduce(lambda i,k:i.get(k), [item] + rule[1].split('.')) if rule[0] == 'rename': del reduce(lambda i,k:i.get(k), [item] + rule[1].split('.')[:-1])[rule[1].split('.')[-1]] yield item
Allow dot notation to map to nested dictionaries
Allow dot notation to map to nested dictionaries
Python
mit
nerevu/riko,nerevu/riko
--- +++ @@ -27,8 +27,11 @@ for item in _INPUT: for rule in rules: - item[rule[2]] = item[rule[1]] + #Map names with dot notation onto nested dictionaries, e.g. 'a.content' -> ['a']['content'] + #todo: optimise by pre-calculating splits + # and if this logic is stable, wrap in util functions and use everywhere items are accessed + reduce(lambda i,k:i.get(k), [item] + rule[2].split('.')[:-1])[rule[2].split('.')[-1]] = reduce(lambda i,k:i.get(k), [item] + rule[1].split('.')) if rule[0] == 'rename': - del item[rule[1]] + del reduce(lambda i,k:i.get(k), [item] + rule[1].split('.')[:-1])[rule[1].split('.')[-1]] yield item