commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
326ae45949c2ee4f53e9c377582313155d9d0b70
kk/models/base.py
kk/models/base.py
from django.conf import settings from django.db import models from django.utils import timezone from django.utils.translation import ugettext_lazy as _ class ModifiableModel(models.Model): created_at = models.DateTimeField(verbose_name=_('Time of creation'), default=timezone.now) created_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Created by'), null=True, blank=True, related_name="%(class)s_created") modified_at = models.DateTimeField(verbose_name=_('Time of modification'), default=timezone.now) modified_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Modified by'), null=True, blank=True, related_name="%(class)s_modified") class Meta: abstract = True
import base64 import struct from django.conf import settings from django.db import models from django.utils import timezone from django.utils.translation import ugettext_lazy as _ def generate_id(): t = time.time() * 1000000 b = base64.b32encode(struct.pack(">Q", int(t)).lstrip(b'\x00')).strip(b'=').lower() return b.decode('utf8') class ModifiableModel(models.Model): id = models.CharField(primary_key=True, max_length=100) created_at = models.DateTimeField(verbose_name=_('Time of creation'), default=timezone.now) created_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Created by'), null=True, blank=True, related_name="%(class)s_created") modified_at = models.DateTimeField(verbose_name=_('Time of modification'), default=timezone.now) modified_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Modified by'), null=True, blank=True, related_name="%(class)s_modified") def save(self, *args, **kwargs): pk_type = self._meta.pk.get_internal_type() if pk_type == 'CharField': if not self.pk: self.pk = generate_id() elif pk_type == 'AutoField': pass else: raise Exception('Unsupported primary key field: %s' % pk_type) super().save(*args, **kwargs) class Meta: abstract = True
Set char primary key. Generate ID.
Set char primary key. Generate ID.
Python
mit
vikoivun/kerrokantasi,stephawe/kerrokantasi,City-of-Helsinki/kerrokantasi,vikoivun/kerrokantasi,stephawe/kerrokantasi,City-of-Helsinki/kerrokantasi,City-of-Helsinki/kerrokantasi,City-of-Helsinki/kerrokantasi,vikoivun/kerrokantasi,stephawe/kerrokantasi
--- +++ @@ -1,16 +1,35 @@ +import base64 +import struct from django.conf import settings from django.db import models from django.utils import timezone from django.utils.translation import ugettext_lazy as _ +def generate_id(): + t = time.time() * 1000000 + b = base64.b32encode(struct.pack(">Q", int(t)).lstrip(b'\x00')).strip(b'=').lower() + return b.decode('utf8') + class ModifiableModel(models.Model): + id = models.CharField(primary_key=True, max_length=100) created_at = models.DateTimeField(verbose_name=_('Time of creation'), default=timezone.now) created_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Created by'), null=True, blank=True, related_name="%(class)s_created") modified_at = models.DateTimeField(verbose_name=_('Time of modification'), default=timezone.now) modified_by = models.ForeignKey(settings.AUTH_USER_MODEL, verbose_name=_('Modified by'), null=True, blank=True, related_name="%(class)s_modified") - + + def save(self, *args, **kwargs): + pk_type = self._meta.pk.get_internal_type() + if pk_type == 'CharField': + if not self.pk: + self.pk = generate_id() + elif pk_type == 'AutoField': + pass + else: + raise Exception('Unsupported primary key field: %s' % pk_type) + super().save(*args, **kwargs) + class Meta: abstract = True
aaf776b94416b63a0da6dfeca6ea04f6fe32d201
systemtests/test_cli.py
systemtests/test_cli.py
import os import os.path from scripttest import TestFileEnvironment def test_persists_data(): env = TestFileEnvironment() # FIXME path to fridge script should be determined in some other way env.run('../../bin/fridge', 'init') env.writefile('somefile', 'with some content') env.run('../../bin/fridge', 'commit') os.unlink(os.path.join(env.base_path, 'somefile')) result = env.run('../../bin/fridge', 'checkout', 'somefile') assert result.files_created['somefile'].bytes == 'with some content'
import os import os.path import scripttest def test_persists_data(): env = scripttest.TestFileEnvironment() # FIXME path to fridge script should be determined in some other way env.run('../../bin/fridge', 'init') env.writefile('somefile', 'with some content') env.run('../../bin/fridge', 'commit') os.unlink(os.path.join(env.base_path, 'somefile')) result = env.run('../../bin/fridge', 'checkout', 'somefile') assert result.files_created['somefile'].bytes == 'with some content'
Fix number of skipped tests when running pytest.
Fix number of skipped tests when running pytest.
Python
mit
jgosmann/fridge,jgosmann/fridge
--- +++ @@ -1,11 +1,11 @@ import os import os.path -from scripttest import TestFileEnvironment +import scripttest def test_persists_data(): - env = TestFileEnvironment() + env = scripttest.TestFileEnvironment() # FIXME path to fridge script should be determined in some other way env.run('../../bin/fridge', 'init') env.writefile('somefile', 'with some content')
b21f540ca7b53aeb569f7034de41da0dc4dd7b03
__init__.py
__init__.py
from vod_metadata.md5_checksum import * from vod_metadata.media_info import * from vod_metadata.parse_config import * from vod_metadata.VodPackage import * (extensions, MediaInfo_path, product, provider_id, prefix, title_category, provider, ecn_2009) = parse_config("./template_values.ini")
import os.path from vod_metadata.md5_checksum import * from vod_metadata.media_info import * from vod_metadata.parse_config import * from vod_metadata.VodPackage import * _script_path = os.path.abspath(__file__) _script_path = os.path.split(_script_path)[0] config_path = os.path.join(_script_path, "template_values.ini") (extensions, MediaInfo_path, product, provider_id, prefix, title_category, provider, ecn_2009) = parse_config(config_path)
Read the configuration file upon import
Read the configuration file upon import
Python
mit
bbayles/vod_metadata
--- +++ @@ -1,7 +1,12 @@ +import os.path from vod_metadata.md5_checksum import * from vod_metadata.media_info import * from vod_metadata.parse_config import * from vod_metadata.VodPackage import * + +_script_path = os.path.abspath(__file__) +_script_path = os.path.split(_script_path)[0] +config_path = os.path.join(_script_path, "template_values.ini") (extensions, MediaInfo_path, @@ -10,4 +15,4 @@ prefix, title_category, provider, - ecn_2009) = parse_config("./template_values.ini") + ecn_2009) = parse_config(config_path)
f691b8d997327a09824881810cf1edaeb53d7579
telemetry/telemetry/internal/backends/app_backend.py
telemetry/telemetry/internal/backends/app_backend.py
# Copyright 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from __future__ import absolute_import import six from py_trace_event import trace_event class AppBackend(six.with_metaclass(trace_event.TracedMetaClass, object)): def __init__(self, app_type, platform_backend): super(AppBackend, self).__init__() self._app = None self._app_type = app_type self._platform_backend = platform_backend def __del__(self): self.Close() def SetApp(self, app): self._app = app @property def app(self): return self._app @property def app_type(self): return self._app_type @property def platform_backend(self): return self._platform_backend def Foreground(self): # TODO(catapult:#2194): Remove the unnecessary pass below when the method # has been implemented on all concrete subclasses. pass # pylint: disable=unnecessary-pass raise NotImplementedError def Background(self): raise NotImplementedError def Close(self): raise NotImplementedError def IsAppRunning(self): raise NotImplementedError
# Copyright 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from __future__ import absolute_import import six from py_trace_event import trace_event class AppBackend(six.with_metaclass(trace_event.TracedMetaClass, object)): def __init__(self, app_type, platform_backend): super(AppBackend, self).__init__() self._app = None self._app_type = app_type self._platform_backend = platform_backend def SetApp(self, app): self._app = app @property def app(self): return self._app @property def app_type(self): return self._app_type @property def platform_backend(self): return self._platform_backend def Foreground(self): # TODO(catapult:#2194): Remove the unnecessary pass below when the method # has been implemented on all concrete subclasses. pass # pylint: disable=unnecessary-pass raise NotImplementedError def Background(self): raise NotImplementedError def Close(self): raise NotImplementedError def IsAppRunning(self): raise NotImplementedError
Remove custom destructor from AppBackend
[Telemetry] Remove custom destructor from AppBackend Closing the backend in the destructor is 1) Redundant since the backend is being closed from Browser.close() 2) Dangerous since it can happen at any moment due to garbage collection and cause deadlocks in tracing code (see the bug). Bug: chromium:1227504 Change-Id: I518eb7153c8929117715834b60bc08c413883866 Reviewed-on: https://chromium-review.googlesource.com/c/catapult/+/3015685 Auto-Submit: Mikhail Khokhlov <26a5474f563e73cb4d4e9a12e954c50f60aae4c8@google.com> Reviewed-by: Wenbin Zhang <45ad8a1ff616a38a2202fd35571df81897c5f14f@google.com> Commit-Queue: Wenbin Zhang <45ad8a1ff616a38a2202fd35571df81897c5f14f@google.com>
Python
bsd-3-clause
catapult-project/catapult,catapult-project/catapult,catapult-project/catapult,catapult-project/catapult,catapult-project/catapult,catapult-project/catapult,catapult-project/catapult
--- +++ @@ -15,9 +15,6 @@ self._app = None self._app_type = app_type self._platform_backend = platform_backend - - def __del__(self): - self.Close() def SetApp(self, app): self._app = app
1fe88c1a619211a93297e1081133017ff2ef0370
scenario/__main__.py
scenario/__main__.py
import sys from scenario import run_scenario def main(args=None): if args is None: args = sys.argv[1:] assert len(args) == 2, 'Usage: python -m foo <executable> <scenario>' executable_path = args[0] scenario_path = args[1] run_scenario(executable_path, scenario_path) if __name__ == '__main__': main()
import sys from scenario import run_scenario def main(args=None): if args is None: args = sys.argv[1:] assert len(args) == 2, 'Usage: scenario <executable> <scenario>' executable_path = args[0] scenario_path = args[1] run_scenario(executable_path, scenario_path) if __name__ == '__main__': main()
Update command line usage with scenario
Update command line usage with scenario
Python
mit
shlomihod/scenario,shlomihod/scenario,shlomihod/scenario
--- +++ @@ -7,7 +7,7 @@ if args is None: args = sys.argv[1:] - assert len(args) == 2, 'Usage: python -m foo <executable> <scenario>' + assert len(args) == 2, 'Usage: scenario <executable> <scenario>' executable_path = args[0] scenario_path = args[1] @@ -16,4 +16,3 @@ if __name__ == '__main__': main() -
c85631d77cd25de520688666a3d0e72537e482eb
acquisition_record.py
acquisition_record.py
""" AcquisitionRecord: database interface class. These classes provide an interface between the database and the top-level ingest algorithm (AbstractIngester and its subclasses). They also provide the implementation of the database and tile store side of the ingest process. They are expected to be independent of the structure of any particular dataset, but will change if the database schema or tile store format changes. """ import logging # Set up logger. LOGGER = logging.getLogger(__name__) LOGGER.setLevel(logging.INFO) class AcquisitionRecord(object): """AcquisitionRecord database interface class.""" def __init__(self, collection, acquisition_id): self.collection = collection self.acquisition_id = acquisition_id def create_dataset_record(self, dataset): pass
""" AcquisitionRecord: database interface class. These classes provide an interface between the database and the top-level ingest algorithm (AbstractIngester and its subclasses). They also provide the implementation of the database and tile store side of the ingest process. They are expected to be independent of the structure of any particular dataset, but will change if the database schema or tile store format changes. Test of github push. """ import logging # Set up logger. LOGGER = logging.getLogger(__name__) LOGGER.setLevel(logging.INFO) class AcquisitionRecord(object): """AcquisitionRecord database interface class.""" def __init__(self, collection, acquisition_id): self.collection = collection self.acquisition_id = acquisition_id def create_dataset_record(self, dataset): pass
Test of github push from Eclipse.
Test of github push from Eclipse.
Python
apache-2.0
ama-jharrison/agdc,GeoscienceAustralia/agdc,jeremyh/agdc,sixy6e/agdc,jeremyh/agdc,smr547/agdc,smr547/agdc,alex-ip/agdc,sixy6e/agdc,GeoscienceAustralia/agdc,alex-ip/agdc,ama-jharrison/agdc
--- +++ @@ -7,6 +7,8 @@ process. They are expected to be independent of the structure of any particular dataset, but will change if the database schema or tile store format changes. + +Test of github push. """ import logging
65734594816b158bdf08b93244795b5dcc8626ba
scrapy/core/downloader/handlers/__init__.py
scrapy/core/downloader/handlers/__init__.py
"""Download handlers for different schemes""" from twisted.internet import defer from scrapy.exceptions import NotSupported, NotConfigured from scrapy.utils.httpobj import urlparse_cached from scrapy.utils.misc import load_object from scrapy import signals class DownloadHandlers(object): def __init__(self, crawler): self._handlers = {} self._notconfigured = {} handlers = crawler.settings.get('DOWNLOAD_HANDLERS_BASE') handlers.update(crawler.settings.get('DOWNLOAD_HANDLERS', {})) for scheme, clspath in handlers.iteritems(): # Allow to disable a handler just like any other # component (extension, middlware, etc). if clspath is None: continue cls = load_object(clspath) try: dh = cls(crawler.settings) except NotConfigured as ex: self._notconfigured[scheme] = str(ex) else: self._handlers[scheme] = dh crawler.signals.connect(self._close, signals.engine_stopped) def download_request(self, request, spider): scheme = urlparse_cached(request).scheme try: handler = self._handlers[scheme].download_request except KeyError: msg = self._notconfigured.get(scheme, \ 'no handler available for that scheme') raise NotSupported("Unsupported URL scheme '%s': %s" % (scheme, msg)) return handler(request, spider) @defer.inlineCallbacks def _close(self, *_a, **_kw): for dh in self._handlers.values(): if hasattr(dh, 'close'): yield dh.close()
"""Download handlers for different schemes""" from twisted.internet import defer from scrapy.exceptions import NotSupported, NotConfigured from scrapy.utils.httpobj import urlparse_cached from scrapy.utils.misc import load_object from scrapy import signals class DownloadHandlers(object): def __init__(self, crawler): self._handlers = {} self._notconfigured = {} handlers = crawler.settings.get('DOWNLOAD_HANDLERS_BASE') handlers.update(crawler.settings.get('DOWNLOAD_HANDLERS', {})) for scheme, clspath in handlers.iteritems(): # Allow to disable a handler just like any other # component (extension, middleware, etc). if clspath is None: continue cls = load_object(clspath) try: dh = cls(crawler.settings) except NotConfigured as ex: self._notconfigured[scheme] = str(ex) else: self._handlers[scheme] = dh crawler.signals.connect(self._close, signals.engine_stopped) def download_request(self, request, spider): scheme = urlparse_cached(request).scheme try: handler = self._handlers[scheme].download_request except KeyError: msg = self._notconfigured.get(scheme, \ 'no handler available for that scheme') raise NotSupported("Unsupported URL scheme '%s': %s" % (scheme, msg)) return handler(request, spider) @defer.inlineCallbacks def _close(self, *_a, **_kw): for dh in self._handlers.values(): if hasattr(dh, 'close'): yield dh.close()
Fix minor typo in DownloaderHandlers comment
Fix minor typo in DownloaderHandlers comment
Python
bsd-3-clause
heamon7/scrapy,redapple/scrapy,Partoo/scrapy,ArturGaspar/scrapy,fpy171/scrapy,AaronTao1990/scrapy,fafaman/scrapy,agreen/scrapy,aivarsk/scrapy,Bourneer/scrapy,w495/scrapy,1yvT0s/scrapy,curita/scrapy,dhenyjarasandy/scrapy,rdowinton/scrapy,Allianzcortex/scrapy,zackslash/scrapy,ashishnerkar1/scrapy,AaronTao1990/scrapy,kmike/scrapy,pfctdayelise/scrapy,olorz/scrapy,KublaikhanGeek/scrapy,agreen/scrapy,tliber/scrapy,kazitanvirahsan/scrapy,farhan0581/scrapy,arush0311/scrapy,heamon7/scrapy,starrify/scrapy,scrapy/scrapy,hwsyy/scrapy,CodeJuan/scrapy,kashyap32/scrapy,jc0n/scrapy,raphaelfruneaux/scrapy,rolando/scrapy,beni55/scrapy,TarasRudnyk/scrapy,hectoruelo/scrapy,lacrazyboy/scrapy,codebhendi/scrapy,rootAvish/scrapy,github-account-because-they-want-it/scrapy,csalazar/scrapy,huoxudong125/scrapy,Geeglee/scrapy,darkrho/scrapy-scrapy,shaform/scrapy,liyy7/scrapy,barraponto/scrapy,foromer4/scrapy,fontenele/scrapy,crasker/scrapy,raphaelfruneaux/scrapy,farhan0581/scrapy,cyberplant/scrapy,wangjun/scrapy,yusofm/scrapy,yarikoptic/scrapy,github-account-because-they-want-it/scrapy,scrapy/scrapy,nowopen/scrapy,kmike/scrapy,URXtech/scrapy,Chenmxs/scrapy,CodeJuan/scrapy,GregoryVigoTorres/scrapy,pombredanne/scrapy,kimimj/scrapy,zjuwangg/scrapy,tagatac/scrapy,z-fork/scrapy,redapple/scrapy,jiezhu2007/scrapy,elacuesta/scrapy,zhangtao11/scrapy,Djlavoy/scrapy,cyrixhero/scrapy,gbirke/scrapy,hbwzhsh/scrapy,Allianzcortex/scrapy,umrashrf/scrapy,rdowinton/scrapy,jorik041/scrapy,sardok/scrapy,ashishnerkar1/scrapy,ArturGaspar/scrapy,darkrho/scrapy-scrapy,avtoritet/scrapy,rdowinton/scrapy,dhenyjarasandy/scrapy,nguyenhongson03/scrapy,cleydson/scrapy,Parlin-Galanodel/scrapy,cyberplant/scrapy,WilliamKinaan/scrapy,Preetwinder/scrapy,Geeglee/scrapy,dgillis/scrapy,nowopen/scrapy,dacjames/scrapy,Ryezhang/scrapy,hyrole/scrapy,yidongliu/scrapy,Allianzcortex/scrapy,Cnfc19932/scrapy,joshlk/scrapy,kashyap32/scrapy,johnardavies/scrapy,eLRuLL/scrapy,codebhendi/scrapy,foromer4/scrapy,TarasRudnyk/scrapy,scorphus/scrapy,godfreyy/scrapy,jamesblunt/scrapy,mgedmin/scrapy,rahulsharma1991/scrapy,cleydson/scrapy,TarasRudnyk/scrapy,IvanGavran/scrapy,webmakin/scrapy,olafdietsche/scrapy,heamon7/scrapy,Bourneer/scrapy,jdemaeyer/scrapy,pablohoffman/scrapy,zhangtao11/scrapy,livepy/scrapy,nguyenhongson03/scrapy,mlyundin/scrapy,Parlin-Galanodel/scrapy,moraesnicol/scrapy,CENDARI/scrapy,dacjames/scrapy,kazitanvirahsan/scrapy,arush0311/scrapy,jiezhu2007/scrapy,curita/scrapy,ssh-odoo/scrapy,fontenele/scrapy,eLRuLL/scrapy,JacobStevenR/scrapy,yarikoptic/scrapy,Adai0808/scrapy-1,pawelmhm/scrapy,beni55/scrapy,rolando-contrib/scrapy,ramiro/scrapy,olorz/scrapy,rootAvish/scrapy,Lucifer-Kim/scrapy,Slater-Victoroff/scrapy,mlyundin/scrapy,smaty1/scrapy,ramiro/scrapy,OpenWhere/scrapy,lacrazyboy/scrapy,wenyu1001/scrapy,gbirke/scrapy,scrapy/scrapy,stenskjaer/scrapy,coderabhishek/scrapy,avtoritet/scrapy,tliber/scrapy,irwinlove/scrapy,Cnfc19932/scrapy,KublaikhanGeek/scrapy,carlosp420/scrapy,finfish/scrapy,jorik041/scrapy,agusc/scrapy,ssh-odoo/scrapy,stenskjaer/scrapy,IvanGavran/scrapy,carlosp420/scrapy,Preetwinder/scrapy,cyberplant/scrapy,CodeJuan/scrapy,wujuguang/scrapy,kalessin/scrapy,crasker/scrapy,zorojean/scrapy,JacobStevenR/scrapy,nguyenhongson03/scrapy,Chenmxs/scrapy,eliasdorneles/scrapy,starrify/scrapy,yarikoptic/scrapy,tntC4stl3/scrapy,legendtkl/scrapy,OpenWhere/scrapy,nfunato/scrapy,legendtkl/scrapy,kalessin/scrapy,finfish/scrapy,hwsyy/scrapy,devGregA/scrapy,ssh-odoo/scrapy,olafdietsche/scrapy,yusofm/scrapy,lacrazyboy/scrapy,godfreyy/scrapy,rklabs/scrapy,Adai0808/scrapy-1,nfunato/scrapy,IvanGavran/scrapy,webmakin/scrapy,dracony/scrapy,wangjun/scrapy,Geeglee/scrapy,elijah513/scrapy,ENjOyAbLE1991/scrapy,fafaman/scrapy,z-fork/scrapy,ramiro/scrapy,snowdream1314/scrapy,joshlk/scrapy,scorphus/scrapy,haiiiiiyun/scrapy,songfj/scrapy,ylcolala/scrapy,pombredanne/scrapy,elijah513/scrapy,zorojean/scrapy,Zephor5/scrapy,redapple/scrapy,rolando-contrib/scrapy,csalazar/scrapy,nikgr95/scrapy,devGregA/scrapy,kimimj/scrapy,ssteo/scrapy,huoxudong125/scrapy,cyrixhero/scrapy,snowdream1314/scrapy,Djlavoy/scrapy,umrashrf/scrapy,JacobStevenR/scrapy,hyrole/scrapy,johnardavies/scrapy,xiao26/scrapy,amboxer21/scrapy,taito/scrapy,haiiiiiyun/scrapy,CENDARI/scrapy,profjrr/scrapy,Digenis/scrapy,codebhendi/scrapy,wzyuliyang/scrapy,cursesun/scrapy,github-account-because-they-want-it/scrapy,pranjalpatil/scrapy,Chenmxs/scrapy,mgedmin/scrapy,legendtkl/scrapy,joshlk/scrapy,godfreyy/scrapy,haiiiiiyun/scrapy,amboxer21/scrapy,dracony/scrapy,Bourneer/scrapy,ENjOyAbLE1991/scrapy,Preetwinder/scrapy,nikgr95/scrapy,Zephor5/scrapy,pfctdayelise/scrapy,liyy7/scrapy,hansenDise/scrapy,snowdream1314/scrapy,dangra/scrapy,jeffreyjinfeng/scrapy,Slater-Victoroff/scrapy,WilliamKinaan/scrapy,profjrr/scrapy,rahulsharma1991/scrapy,bmess/scrapy,taito/scrapy,URXtech/scrapy,zjuwangg/scrapy,yidongliu/scrapy,eLRuLL/scrapy,chekunkov/scrapy,hbwzhsh/scrapy,fqul/scrapy,fqul/scrapy,z-fork/scrapy,finfish/scrapy,stenskjaer/scrapy,darkrho/scrapy-scrapy,kashyap32/scrapy,Timeship/scrapy,tntC4stl3/scrapy,pranjalpatil/scrapy,livepy/scrapy,agusc/scrapy,yidongliu/scrapy,fpy171/scrapy,xiao26/scrapy,devGregA/scrapy,Digenis/scrapy,tagatac/scrapy,pranjalpatil/scrapy,elacuesta/scrapy,wujuguang/scrapy,cyrixhero/scrapy,OpenWhere/scrapy,olafdietsche/scrapy,aivarsk/scrapy,ylcolala/scrapy,Parlin-Galanodel/scrapy,jc0n/scrapy,nowopen/scrapy,kalessin/scrapy,wzyuliyang/scrapy,hectoruelo/scrapy,webmakin/scrapy,farhan0581/scrapy,barraponto/scrapy,nikgr95/scrapy,hansenDise/scrapy,yusofm/scrapy,bmess/scrapy,irwinlove/scrapy,livepy/scrapy,starrify/scrapy,sigma-random/scrapy,dangra/scrapy,rolando/scrapy,1yvT0s/scrapy,hbwzhsh/scrapy,dacjames/scrapy,URXtech/scrapy,CENDARI/scrapy,Lucifer-Kim/scrapy,beni55/scrapy,umrashrf/scrapy,rootAvish/scrapy,smaty1/scrapy,hyrole/scrapy,mgedmin/scrapy,ssteo/scrapy,dangra/scrapy,Timeship/scrapy,KublaikhanGeek/scrapy,sigma-random/scrapy,mlyundin/scrapy,irwinlove/scrapy,coderabhishek/scrapy,cursesun/scrapy,Ryezhang/scrapy,Digenis/scrapy,dhenyjarasandy/scrapy,elacuesta/scrapy,rklabs/scrapy,rolando-contrib/scrapy,Cnfc19932/scrapy,Slater-Victoroff/scrapy,profjrr/scrapy,AaronTao1990/scrapy,jdemaeyer/scrapy,elijah513/scrapy,Ryezhang/scrapy,famorted/scrapy,Partoo/scrapy,pawelmhm/scrapy,Zephor5/scrapy,liyy7/scrapy,csalazar/scrapy,wenyu1001/scrapy,dgillis/scrapy,eliasdorneles/scrapy,hansenDise/scrapy,crasker/scrapy,raphaelfruneaux/scrapy,smaty1/scrapy,songfj/scrapy,1yvT0s/scrapy,aivarsk/scrapy,Partoo/scrapy,kazitanvirahsan/scrapy,carlosp420/scrapy,fafaman/scrapy,hwsyy/scrapy,olorz/scrapy,wenyu1001/scrapy,ENjOyAbLE1991/scrapy,avtoritet/scrapy,Timeship/scrapy,chekunkov/scrapy,amboxer21/scrapy,kimimj/scrapy,pombredanne/scrapy,dracony/scrapy,fqul/scrapy,tliber/scrapy,WilliamKinaan/scrapy,fontenele/scrapy,pfctdayelise/scrapy,rahulsharma1991/scrapy,Lucifer-Kim/scrapy,fpy171/scrapy,shaform/scrapy,jeffreyjinfeng/scrapy,w495/scrapy,jdemaeyer/scrapy,jiezhu2007/scrapy,zhangtao11/scrapy,ylcolala/scrapy,famorted/scrapy,rolando/scrapy,bmess/scrapy,pawelmhm/scrapy,wangjun/scrapy,YeelerG/scrapy,GregoryVigoTorres/scrapy,wujuguang/scrapy,scorphus/scrapy,hectoruelo/scrapy,w495/scrapy,agreen/scrapy,tntC4stl3/scrapy,YeelerG/scrapy,chekunkov/scrapy,Adai0808/scrapy-1,taito/scrapy,zackslash/scrapy,jorik041/scrapy,pablohoffman/scrapy,zorojean/scrapy,cleydson/scrapy,kmike/scrapy,jc0n/scrapy,zjuwangg/scrapy,dgillis/scrapy,pablohoffman/scrapy,ArturGaspar/scrapy,tagatac/scrapy,arush0311/scrapy,eliasdorneles/scrapy,xiao26/scrapy,agusc/scrapy,johnardavies/scrapy,sardok/scrapy,moraesnicol/scrapy,ssteo/scrapy,wzyuliyang/scrapy,songfj/scrapy,foromer4/scrapy,famorted/scrapy,rklabs/scrapy,YeelerG/scrapy,coderabhishek/scrapy,zackslash/scrapy,curita/scrapy,barraponto/scrapy,GregoryVigoTorres/scrapy,moraesnicol/scrapy,cursesun/scrapy,jeffreyjinfeng/scrapy,nfunato/scrapy,Djlavoy/scrapy,jamesblunt/scrapy,huoxudong125/scrapy,shaform/scrapy
--- +++ @@ -16,7 +16,7 @@ handlers.update(crawler.settings.get('DOWNLOAD_HANDLERS', {})) for scheme, clspath in handlers.iteritems(): # Allow to disable a handler just like any other - # component (extension, middlware, etc). + # component (extension, middleware, etc). if clspath is None: continue cls = load_object(clspath)
d6bfc8be7944bd8495a21d9db065990148e6c466
tests/template_error.py
tests/template_error.py
from docxtpl import DocxTemplate, RichText from jinja2.exceptions import TemplateError import six six.print_('=' * 80) six.print_("Generating template error for testing (so it is safe to ignore) :") six.print_('.' * 80) try: tpl = DocxTemplate('test_files/template_error_tpl.docx') tpl.render({ 'test_variable' : 'test variable value' }) except TemplateError as the_error: six.print_(six.text_type(the_error)) if hasattr(the_error, 'docx_context'): print "Context:" for line in the_error.docx_context: six.print_(line) tpl.save('test_files/template_error.docx') six.print_('.' * 80) six.print_(" End of TemplateError Test ") six.print_('=' * 80)
from docxtpl import DocxTemplate, RichText from jinja2.exceptions import TemplateError import six six.print_('=' * 80) six.print_("Generating template error for testing (so it is safe to ignore) :") six.print_('.' * 80) try: tpl = DocxTemplate('test_files/template_error_tpl.docx') tpl.render({ 'test_variable' : 'test variable value' }) except TemplateError as the_error: six.print_(six.text_type(the_error)) if hasattr(the_error, 'docx_context'): six.print_("Context:") for line in the_error.docx_context: six.print_(line) tpl.save('test_files/template_error.docx') six.print_('.' * 80) six.print_(" End of TemplateError Test ") six.print_('=' * 80)
Fix test incompatibility with Python 3 versions
Fix test incompatibility with Python 3 versions Replaced 'print' instruction with call of a 'six' package's implementation compatible with Python 2 as well as Python 3.
Python
lgpl-2.1
elapouya/python-docx-template
--- +++ @@ -13,7 +13,7 @@ except TemplateError as the_error: six.print_(six.text_type(the_error)) if hasattr(the_error, 'docx_context'): - print "Context:" + six.print_("Context:") for line in the_error.docx_context: six.print_(line) tpl.save('test_files/template_error.docx')
6403229da220fddac236a8e3ccf061446c37e27c
auditlog/__openerp__.py
auditlog/__openerp__.py
# -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2013 ABF OSIELL (<http://osiell.com>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## { 'name': "Audit Log", 'version': "1.0", 'author': "ABF OSIELL", 'website': "http://www.osiell.com", 'category': "Tools", 'depends': [ 'base', ], 'data': [ 'security/ir.model.access.csv', 'views/auditlog_view.xml', ], 'application': True, 'installable': True, 'pre_init_hook': 'pre_init_hook', }
# -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2013 ABF OSIELL (<http://osiell.com>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## { 'name': "Audit Log", 'version': "1.0", 'author': "ABF OSIELL,Odoo Community Association (OCA)", 'website': "http://www.osiell.com", 'category': "Tools", 'depends': [ 'base', ], 'data': [ 'security/ir.model.access.csv', 'views/auditlog_view.xml', ], 'application': True, 'installable': True, 'pre_init_hook': 'pre_init_hook', }
Add OCA as author of OCA addons
Add OCA as author of OCA addons In order to get visibility on https://www.odoo.com/apps the OCA board has decided to add the OCA as author of all the addons maintained as part of the association.
Python
agpl-3.0
Vauxoo/server-tools,Vauxoo/server-tools,Vauxoo/server-tools
--- +++ @@ -22,7 +22,7 @@ { 'name': "Audit Log", 'version': "1.0", - 'author': "ABF OSIELL", + 'author': "ABF OSIELL,Odoo Community Association (OCA)", 'website': "http://www.osiell.com", 'category': "Tools", 'depends': [
33394f4081880c2718f1c017fb90588628c2bfcc
tests/test_extension.py
tests/test_extension.py
import unittest import mock from mopidy_spotify import Extension, backend as backend_lib class ExtensionTest(unittest.TestCase): def test_get_default_config(self): ext = Extension() config = ext.get_default_config() self.assertIn('[spotify]', config) self.assertIn('enabled = true', config) def test_get_config_schema(self): ext = Extension() schema = ext.get_config_schema() self.assertIn('username', schema) self.assertIn('password', schema) self.assertIn('bitrate', schema) self.assertIn('timeout', schema) self.assertIn('cache_dir', schema) def test_setup(self): registry = mock.Mock() ext = Extension() ext.setup(registry) registry.add.assert_called_with('backend', backend_lib.SpotifyBackend)
import mock from mopidy_spotify import Extension, backend as backend_lib def test_get_default_config(): ext = Extension() config = ext.get_default_config() assert '[spotify]' in config assert 'enabled = true' in config def test_get_config_schema(): ext = Extension() schema = ext.get_config_schema() assert 'username' in schema assert 'password' in schema assert 'bitrate' in schema assert 'timeout' in schema assert 'cache_dir' in schema def test_setup(): registry = mock.Mock() ext = Extension() ext.setup(registry) registry.add.assert_called_with('backend', backend_lib.SpotifyBackend)
Convert extension tests to pytest syntax
tests: Convert extension tests to pytest syntax
Python
apache-2.0
jodal/mopidy-spotify,kingosticks/mopidy-spotify,mopidy/mopidy-spotify
--- +++ @@ -1,35 +1,33 @@ -import unittest - import mock from mopidy_spotify import Extension, backend as backend_lib -class ExtensionTest(unittest.TestCase): +def test_get_default_config(): + ext = Extension() - def test_get_default_config(self): - ext = Extension() + config = ext.get_default_config() - config = ext.get_default_config() + assert '[spotify]' in config + assert 'enabled = true' in config - self.assertIn('[spotify]', config) - self.assertIn('enabled = true', config) - def test_get_config_schema(self): - ext = Extension() +def test_get_config_schema(): + ext = Extension() - schema = ext.get_config_schema() + schema = ext.get_config_schema() - self.assertIn('username', schema) - self.assertIn('password', schema) - self.assertIn('bitrate', schema) - self.assertIn('timeout', schema) - self.assertIn('cache_dir', schema) + assert 'username' in schema + assert 'password' in schema + assert 'bitrate' in schema + assert 'timeout' in schema + assert 'cache_dir' in schema - def test_setup(self): - registry = mock.Mock() - ext = Extension() - ext.setup(registry) +def test_setup(): + registry = mock.Mock() - registry.add.assert_called_with('backend', backend_lib.SpotifyBackend) + ext = Extension() + ext.setup(registry) + + registry.add.assert_called_with('backend', backend_lib.SpotifyBackend)
69d2620ee64d367331edcf0260c73034384aae8e
subprocrunner/retry.py
subprocrunner/retry.py
import time from random import uniform from typing import Callable, Optional class Retry: def __init__(self, total: int = 3, backoff_factor: float = 0.2, jitter: float = 0.2) -> None: self.total = total self.__backoff_factor = backoff_factor self.__jitter = jitter if self.total <= 0: raise ValueError("total must be greater than zero") if self.__backoff_factor <= 0: raise ValueError("backoff_factor must be greater than zero") if self.__jitter <= 0: raise ValueError("jitter must be greater than zero") def calc_backoff_time(self, attempt: int) -> float: sleep_duration = self.__backoff_factor * (2 ** max(0, attempt - 1)) sleep_duration += uniform(0.5 * self.__jitter, 1.5 * self.__jitter) return sleep_duration def sleep_before_retry( self, attempt: int, logging_method: Optional[Callable] = None, retry_target: Optional[str] = None, ) -> float: sleep_duration = self.calc_backoff_time(attempt) if logging_method: if retry_target: msg = "Retrying '{}' in ".format(retry_target) else: msg = "Retrying in " msg += "{:.2f} seconds ... (attempt={})".format(sleep_duration, attempt) logging_method(msg) time.sleep(sleep_duration) return sleep_duration
import time from random import uniform from typing import Callable, Optional class Retry: def __init__( self, total: int = 3, backoff_factor: float = 0.2, jitter: float = 0.2, quiet: bool = False ) -> None: self.total = total self.__backoff_factor = backoff_factor self.__jitter = jitter self.__quiet = quiet if self.total <= 0: raise ValueError("total must be greater than zero") if self.__backoff_factor <= 0: raise ValueError("backoff_factor must be greater than zero") if self.__jitter <= 0: raise ValueError("jitter must be greater than zero") def calc_backoff_time(self, attempt: int) -> float: sleep_duration = self.__backoff_factor * (2 ** max(0, attempt - 1)) sleep_duration += uniform(0.5 * self.__jitter, 1.5 * self.__jitter) return sleep_duration def sleep_before_retry( self, attempt: int, logging_method: Optional[Callable] = None, retry_target: Optional[str] = None, ) -> float: sleep_duration = self.calc_backoff_time(attempt) if logging_method and not self.__quiet: if retry_target: msg = "Retrying '{}' in ".format(retry_target) else: msg = "Retrying in " msg += "{:.2f} seconds ... (attempt={})".format(sleep_duration, attempt) logging_method(msg) time.sleep(sleep_duration) return sleep_duration
Add quiet mode support for Retry
Add quiet mode support for Retry
Python
mit
thombashi/subprocrunner,thombashi/subprocrunner
--- +++ @@ -4,10 +4,13 @@ class Retry: - def __init__(self, total: int = 3, backoff_factor: float = 0.2, jitter: float = 0.2) -> None: + def __init__( + self, total: int = 3, backoff_factor: float = 0.2, jitter: float = 0.2, quiet: bool = False + ) -> None: self.total = total self.__backoff_factor = backoff_factor self.__jitter = jitter + self.__quiet = quiet if self.total <= 0: raise ValueError("total must be greater than zero") @@ -32,7 +35,7 @@ ) -> float: sleep_duration = self.calc_backoff_time(attempt) - if logging_method: + if logging_method and not self.__quiet: if retry_target: msg = "Retrying '{}' in ".format(retry_target) else:
d676065cb9f137c5feb18b125d0d30dfae4e0b65
Dice.py
Dice.py
import random class Die(object): def __init__(self, sides = 6): self.sides = sides self.held = False self.die_face = 1 def change_held(self, held): self.held = held def roll_die(self): if (self.held == False): self.die_face = random.randint(1, self.sides) else: pass def get_die_face(self): return self.die_face class DiceBag(object): def __init__(self): self.dice = [] self.dice_roll = [] def add_die_obj(self, die_obj): self.dice.append(die_obj) def remove_die(self, die_obj): self.dice.remove(die_obj) def remove_die_index(self, index): del self.dice[index] def add_die_notation(self, standard_die_notation): lst_notation = standard_die_notation.split("d") for i in int(lst_notation[0]): die1 = Die(int(lst_notation[1])) self.dice.append(die1) def roll_all(self): for obj in self.dice: obj.roll_die() self.dice_roll.append(obj.get_die_face())
import random class Die(object): def __init__(self, sides = 6): self.sides = sides self.held = False self.die_face = 1 def change_held(self, held): self.held = held def roll_die(self): if (self.held == False): self.die_face = random.randint(1, self.sides) else: pass def get_die_face(self): return self.die_face class DiceBag(object): def __init__(self): self.dice = [] self.dice_roll = [] def add_die_obj(self, die_obj): self.dice.append(die_obj) def remove_die(self, die_obj): self.dice.remove(die_obj) def remove_die_index(self, index): del self.dice[index] def add_die_notation(self, standard_die_notation): lst_notation = standard_die_notation.split("d") for i in int(lst_notation[0]): die1 = Die(int(lst_notation[1])) self.dice.append(die1) def roll_all(self): for obj in self.dice: obj.roll_die() self.dice_roll.append(obj.get_die_face()) def get_dice_roll(self): return self.dice_roll
Add get dice roll function
Add get dice roll function
Python
mit
achyutreddy24/DiceGame
--- +++ @@ -33,3 +33,5 @@ for obj in self.dice: obj.roll_die() self.dice_roll.append(obj.get_die_face()) + def get_dice_roll(self): + return self.dice_roll
8ea896e3290d441e6025822cc4e67b2fd86c3a8c
social_django/compat.py
social_django/compat.py
# coding=utf-8 import six import django from django.db import models try: from django.urls import reverse except ImportError: from django.core.urlresolvers import reverse if django.VERSION >= (1, 10): from django.utils.deprecation import MiddlewareMixin else: MiddlewareMixin = object def get_rel_model(field): if django.VERSION >= (2, 0): return field.remote_field.model user_model = field.rel.to if isinstance(user_model, six.string_types): app_label, model_name = user_model.split('.') user_model = models.get_model(app_label, model_name) return user_model def get_request_port(request): if django.VERSION >= (1, 9): return request.get_port() host_parts = request.get_host().partition(':') return host_parts[2] or request.META['SERVER_PORT']
# coding=utf-8 import six import django from django.db import models try: from django.urls import reverse except ImportError: from django.core.urlresolvers import reverse try: from django.utils.deprecation import MiddlewareMixin except ImportError: MiddlewareMixin = object def get_rel_model(field): if django.VERSION >= (2, 0): return field.remote_field.model user_model = field.rel.to if isinstance(user_model, six.string_types): app_label, model_name = user_model.split('.') user_model = models.get_model(app_label, model_name) return user_model def get_request_port(request): if django.VERSION >= (1, 9): return request.get_port() host_parts = request.get_host().partition(':') return host_parts[2] or request.META['SERVER_PORT']
Remove version check in favor of import error check
Remove version check in favor of import error check
Python
bsd-3-clause
python-social-auth/social-app-django,python-social-auth/social-app-django,python-social-auth/social-app-django
--- +++ @@ -2,15 +2,16 @@ import six import django from django.db import models + try: from django.urls import reverse except ImportError: from django.core.urlresolvers import reverse -if django.VERSION >= (1, 10): +try: from django.utils.deprecation import MiddlewareMixin -else: +except ImportError: MiddlewareMixin = object
342f2a948ba88d6c67c003457923b135234088a0
JSON.py
JSON.py
import os import json import ast from functions import quit from encryption import encrypt, decrypt global name global key def setOfflineUsername(_name, _key): global name global key name = _name key = _key def getServicesOffline(): dir_path = os.path.expanduser("~/.passman") file_path = os.path.expanduser("~/.passman/{}.json".format(name)) if not os.path.isfile(file_path) or \ not os.path.isdir(dir_path): print("No local file found - exiting") quit() with open(file_path) as data_file: data = data_file.read() data = ast.literal_eval(data)['data'] for service in data: service['service'] = decrypt(service['service'], key) service['serviceUserName'] = decrypt(service['serviceUserName'], key) service['servicePassword'] = decrypt(service['servicePassword'], key) service['serviceUrl'] = decrypt(service['serviceUrl'], key) return data def getServiceDataOffline(sname): global name serviceArray = getServicesOffline() for service in serviceArray: if service['service'] == sname: return service return False
import os import json import ast from functions import quit from encryption import encrypt, decrypt global name global key def setOfflineUsername(_name, _key): global name global key name = _name key = _key def getServicesOffline(): global name dir_path = os.path.expanduser("~/.passman") file_path = os.path.expanduser("~/.passman/{}.json".format(name)) if not os.path.isfile(file_path) or \ not os.path.isdir(dir_path): print("No local file found - exiting") quit() with open(file_path) as data_file: data = data_file.read() data = ast.literal_eval(data)['data'] for service in data: service['service'] = decrypt(service['service'], key) service['serviceUserName'] = decrypt(service['serviceUserName'], key) service['servicePassword'] = decrypt(service['servicePassword'], key) service['serviceUrl'] = decrypt(service['serviceUrl'], key) return data def getServiceDataOffline(sname): global name serviceArray = getServicesOffline() for service in serviceArray: if service['service'] == sname: return service return False
Call global name in getservicesoffline function
Call global name in getservicesoffline function
Python
mit
regexpressyourself/passman
--- +++ @@ -12,9 +12,8 @@ name = _name key = _key - - def getServicesOffline(): + global name dir_path = os.path.expanduser("~/.passman") file_path = os.path.expanduser("~/.passman/{}.json".format(name)) if not os.path.isfile(file_path) or \
a6ae05c13666b83a1f1a8707fe21972bd1f758d9
walltime.py
walltime.py
#!/usr/bin/env python """ Created on Fri Mar 14 15:25:36 2014 @author: ibackus """ import matplotlib.pyplot as plt import numpy as np import datetime import sys if len(sys.argv) < 2: print 'USAGE: walltime filename' else: fname = sys.argv[-1] log_file = np.genfromtxt(fname, comments='#', delimiter=' ') walltime_total = datetime.timedelta(seconds = log_file[:,-1].sum()) walltime_avg = datetime.timedelta(seconds = log_file[:,-1].mean()) print 'Total walltime: ' print str(walltime_total) print 'Average walltime per step:' print str(walltime_avg) plt.plot(log_file[:,-1],'x') plt.show()
#!/usr/bin/env python """ Created on Fri Mar 14 15:25:36 2014 @author: ibackus """ import time t0 = time.time() import matplotlib.pyplot as plt import numpy as np import datetime import sys t1 = time.time() print 'Importing took {} s'.format(t1-t0) if len(sys.argv) < 2: print 'USAGE: walltime filename' else: fname = sys.argv[-1] log_file = np.genfromtxt(fname, comments='#', delimiter=' ') walltime_total = datetime.timedelta(seconds = log_file[:,-1].sum()) walltime_avg = datetime.timedelta(seconds = log_file[:,-1].mean()) print 'Total walltime: ' print str(walltime_total) print 'Average walltime per step:' print str(walltime_avg) plt.plot(log_file[:,-1],'x') t2 = time.time() print 'Running took an extra {} s'.format(t2-t1) print 'For a total of {} s'.format(t2 - t0) plt.show()
Print statements added for profiling
Print statements added for profiling
Python
mit
ibackus/custom_python_packages,trquinn/custom_python_packages
--- +++ @@ -5,10 +5,16 @@ @author: ibackus """ +import time + +t0 = time.time() import matplotlib.pyplot as plt import numpy as np import datetime import sys + +t1 = time.time() +print 'Importing took {} s'.format(t1-t0) if len(sys.argv) < 2: @@ -27,4 +33,9 @@ print str(walltime_avg) plt.plot(log_file[:,-1],'x') + + t2 = time.time() + print 'Running took an extra {} s'.format(t2-t1) + print 'For a total of {} s'.format(t2 - t0) + plt.show()
692a6d4480e917ff2648bac7ac4975f981e4c571
scripts/util/assignCounty.py
scripts/util/assignCounty.py
from pyIEM import iemdb import re i = iemdb.iemdb() mydb = i["mesosite"] rs = mydb.query("select s.id, c.name from stations s, counties c WHERE \ s.geom && c.the_geom and s.county IS NULL").dictresult() for i in range(len(rs)): id = rs[i]['id'] cnty = re.sub("'", " ", rs[i]['name']) print id, cnty mydb.query("UPDATE stations SET county = '%s' WHERE id = '%s'" \ % (cnty, id) )
from pyIEM import iemdb import re i = iemdb.iemdb() mydb = i["mesosite"] rs = mydb.query(""" select s.id, c.name from stations s, counties c, states t WHERE ST_Contains(c.the_geom, s.geom) and s.geom && c.the_geom and s.county IS NULL and s.state = t.state_abbr and t.state_fips = c.state_fips """).dictresult() for i in range(len(rs)): id = rs[i]['id'] cnty = re.sub("'", " ", rs[i]['name']) print id, cnty mydb.query("UPDATE stations SET county = '%s' WHERE id = '%s'" \ % (cnty, id) )
Make sure that the county is in the right state even.
Make sure that the county is in the right state even.
Python
mit
akrherz/iem,akrherz/iem,akrherz/iem,akrherz/iem,akrherz/iem
--- +++ @@ -4,8 +4,12 @@ i = iemdb.iemdb() mydb = i["mesosite"] -rs = mydb.query("select s.id, c.name from stations s, counties c WHERE \ - s.geom && c.the_geom and s.county IS NULL").dictresult() +rs = mydb.query(""" + select s.id, c.name from stations s, counties c, states t WHERE + ST_Contains(c.the_geom, s.geom) and s.geom && c.the_geom + and s.county IS NULL and s.state = t.state_abbr and + t.state_fips = c.state_fips +""").dictresult() for i in range(len(rs)): id = rs[i]['id']
2dbd2d385e821cee9a8bc8414bfba71c8b4dbc06
tests/test_ehrcorral.py
tests/test_ehrcorral.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_ehrcorral ---------------------------------- Tests for `ehrcorral` module. """ import unittest from ehrcorral import ehrcorral class TestEhrcorral(unittest.TestCase): def setUp(self): pass def test_something(self): pass def tearDown(self): pass
#!/usr/bin/env python # -*- coding: utf-8 -*- """ test_ehrcorral ---------------------------------- Tests for `ehrcorral` module. """ from __future__ import print_function from __future__ import division from __future__ import absolute_import from __future__ import unicode_literals import unittest from ehrcorral import ehrcorral from faker import Faker fake = Faker() fake.seed(8548) class TestEHRcorral(unittest.TestCase): def setUp(self): profile_fields = ['name', 'birthdate', 'ssn', 'address'] self.herd = [fake.profile(fields=profile_fields) for n in xrange(100)] def test_something(self): pass def tearDown(self): pass
Add test case setUp to generate fake patient info
Add test case setUp to generate fake patient info
Python
isc
nsh87/ehrcorral
--- +++ @@ -8,15 +8,25 @@ Tests for `ehrcorral` module. """ +from __future__ import print_function +from __future__ import division +from __future__ import absolute_import +from __future__ import unicode_literals + import unittest from ehrcorral import ehrcorral +from faker import Faker + +fake = Faker() +fake.seed(8548) -class TestEhrcorral(unittest.TestCase): +class TestEHRcorral(unittest.TestCase): def setUp(self): - pass + profile_fields = ['name', 'birthdate', 'ssn', 'address'] + self.herd = [fake.profile(fields=profile_fields) for n in xrange(100)] def test_something(self): pass
4783f7047500865da06202a7d6d777801cf49c71
Box.py
Box.py
class Box: def __init__(self, length, width, height): self.length = length self.width = width self.height = height self.plist = list() def main(): N = input() box = list() for i in range(N): x = input() x = x.split('') b = Box(x[0], x[1], x[2]) box.append(b) if __name__ == "__main__": main()
class Box: def __init__(self, length, width, height): self.length = length self.width = width self.height = height self.plist = list() self.plength = 0 def __lt__(self, other): return (self.length < other.length and self.width < other.width and self.height < other.height) def link_to(self, box): self.plist.append(box) self.plength += 1 """ Test for input: def print_content(self): print(self.length, self.width, self.height) """ def main(): N = int(input()) box = list() for i in range(N): x = input() x = x.split(' ') x = Box(x[0], x[1], x[2]) box.append(x) # Test: # for i, x in enumerate(box): # x.print_content() if __name__ == "__main__": main()
Redefine function __lt__ and define a function to set link with others.
Redefine function __lt__ and define a function to set link with others.
Python
mit
hane1818/Algorithm_HW4_box_problem
--- +++ @@ -4,16 +4,34 @@ self.width = width self.height = height self.plist = list() + self.plength = 0 + + def __lt__(self, other): + return (self.length < other.length + and self.width < other.width + and self.height < other.height) + + def link_to(self, box): + self.plist.append(box) + self.plength += 1 +""" + Test for input: + def print_content(self): + print(self.length, self.width, self.height) +""" def main(): - N = input() + N = int(input()) box = list() for i in range(N): x = input() - x = x.split('') - b = Box(x[0], x[1], x[2]) - box.append(b) + x = x.split(' ') + x = Box(x[0], x[1], x[2]) + box.append(x) +# Test: +# for i, x in enumerate(box): +# x.print_content() if __name__ == "__main__": main()
ea15b51ad444eeca3fdbc9eeb30fb8434ec3bfbd
pyscores/api_wrapper.py
pyscores/api_wrapper.py
import json import os import requests class APIWrapper(object): def __init__(self, base_url=None, auth_token=None): if base_url: self.base_url = base_url else: self.base_url = "http://api.football-data.org/v1" if auth_token: self.headers = { 'X-Auth-Token': auth_token } else: self.headers = {} def do_request(self, url, filters=None): params = filters if filters else {} r = requests.get(url=url, params=params, headers=self.headers) if r.status_code == requests.codes.ok: return r.json() return None def all_competitions(self): url = "%s/competitions" % self.base_url response = self.do_request(url=url) return response def main(): api = APIWrapper(auth_token=os.environ["PYSCORES_KEY"]) res = api.do_request("http://api.football-data.org/v1/competitions") print(res) if __name__ == "__main__": main()
import os import requests class APIWrapper(object): def __init__(self, base_url=None, auth_token=None): if base_url: self.base_url = base_url else: self.base_url = "http://api.football-data.org/v1" if auth_token: self.headers = { 'X-Auth-Token': auth_token } else: self.headers = {} def do_request(self, url, filters=None): params = filters if filters else {} r = requests.get(url=url, params=params, headers=self.headers) if r.status_code == requests.codes.ok: return r.json() return None def all_competitions(self): url = "%s/competitions" % self.base_url response = self.do_request(url=url) return response def main(): api = APIWrapper(auth_token=os.environ["PYSCORES_KEY"]) res = api.do_request("http://api.football-data.org/v1/competitions") print(res) if __name__ == "__main__": main()
Remove unused json import in api wrapper
Remove unused json import in api wrapper
Python
mit
conormag94/pyscores
--- +++ @@ -1,4 +1,3 @@ -import json import os import requests
3d48732d577514d888ba5769a27d811d55fd9979
app.py
app.py
from flask import Flask import subprocess from config import repos app = Flask(__name__) @app.route("/", methods=['GET']) def hello(): current_repo = repos.get('key') remote_name = current_repo('remote_name') remote_branch = current_repo('remote_branch') local_dir = current_repo('local_dir') cmd = ["cd %s && git reset --hard && git pull %s %s" % (local_dir, remote_name, remote_branch),""] p = subprocess.Popen(cmd, shell=True, close_fds=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) out,err = p.communicate() return out if __name__ == "__main__": app.run(host='0.0.0.0')
from flask import Flask from flask import request import subprocess from config import repos app = Flask(__name__) @app.route("/", methods=['GET']) def hello(): repo_id = request.args.get('key') current_repo = repos.get(repo_id) remote_name = current_repo.get('remote_name') remote_branch = current_repo.get('remote_branch') local_dir = current_repo.get('local_dir') cmd = ["cd %s && git reset --hard && git pull %s %s" % (local_dir, remote_name, remote_branch),""] p = subprocess.Popen(cmd, shell=True, close_fds=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) out,err = p.communicate() return out if __name__ == "__main__": app.run(host='0.0.0.0')
Send repo id in get param
Send repo id in get param
Python
mit
Heads-and-Hands/pullover
--- +++ @@ -1,4 +1,5 @@ from flask import Flask +from flask import request import subprocess from config import repos @@ -6,10 +7,11 @@ @app.route("/", methods=['GET']) def hello(): - current_repo = repos.get('key') - remote_name = current_repo('remote_name') - remote_branch = current_repo('remote_branch') - local_dir = current_repo('local_dir') + repo_id = request.args.get('key') + current_repo = repos.get(repo_id) + remote_name = current_repo.get('remote_name') + remote_branch = current_repo.get('remote_branch') + local_dir = current_repo.get('local_dir') cmd = ["cd %s && git reset --hard && git pull %s %s" % (local_dir, remote_name, remote_branch),""] p = subprocess.Popen(cmd, shell=True, close_fds=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
69a294d2a7aeab592dfa08e42423d8741aeb3828
app.py
app.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from random import randint from flask import Flask, request, render_template, url_for, redirect from pyhipku import encode app = Flask(__name__) @app.route('/<current_ip>') def index(current_ip): your_ip = request.remote_addr lines = encode(current_ip).split('\n') return render_template('cover.html', lines=lines, your_ip=your_ip, current_ip=current_ip) @app.route('/') def get_ip(): return redirect(url_for('index', current_ip=request.remote_addr)) @app.route('/random') def random_ip(): random_ip = '.'.join(map(str, [randint(0, 255) for _ in range(4)])) return redirect(url_for('index', current_ip=random_ip)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
#!/usr/bin/env python # -*- coding: utf-8 -*- import random from flask import Flask, request, render_template, url_for, redirect from pyhipku import encode app = Flask(__name__) @app.route('/<current_ip>') def index(current_ip): your_ip = request.remote_addr lines = encode(current_ip).split('\n') return render_template('cover.html', lines=lines, your_ip=your_ip, current_ip=current_ip) @app.route('/') def get_ip(): return redirect(url_for('index', current_ip=request.remote_addr)) @app.route('/random') def random_ip(): if random.random() < 0.5: random_ip = '.'.join(map(str, [random.randint(0, 255) for _ in range(4)])) else: random_ip = ':'.join('{0:x}'.format(random.randint(0,2**16-1)) for _ in range(8)) return redirect(url_for('index', current_ip=random_ip)) if __name__ == '__main__': app.run(debug=True, host='0.0.0.0')
Support IPv6 in random ip
Support IPv6 in random ip
Python
mit
lord63/pyhipku_web,lord63/pyhipku_web
--- +++ @@ -1,7 +1,7 @@ #!/usr/bin/env python # -*- coding: utf-8 -*- -from random import randint +import random from flask import Flask, request, render_template, url_for, redirect from pyhipku import encode @@ -25,7 +25,12 @@ @app.route('/random') def random_ip(): - random_ip = '.'.join(map(str, [randint(0, 255) for _ in range(4)])) + if random.random() < 0.5: + random_ip = '.'.join(map(str, + [random.randint(0, 255) for _ in range(4)])) + else: + random_ip = ':'.join('{0:x}'.format(random.randint(0,2**16-1)) + for _ in range(8)) return redirect(url_for('index', current_ip=random_ip))
552ea94c9fe2a42a8041d986e929b7defcdc4a4e
bot.py
bot.py
#! /usr/bin/env python from time import gmtime, strftime from foaas import foaas from diaspy_client import Client import re import urllib2 client = Client() notify = client.notifications() for n in notify: if not n.unread: continue idm = re.search('href=\\"/posts/(\d+?)\\"', n._data['note_html']) if hasattr(idm, 'group'): post_id = idm.group(1) text = client.getPostText(post_id) m = re.search('^\s*\@\{[^\};]+;[^\};]+\}\s+(\/.+?)$', text) if hasattr(m, 'group'): try: command = m.group(1) client.post(foaas(command)) except urllib2.URLError: client.comment(post_id, "Fuck this! Something went wrong :\\") else: client.comment(post_id, "Fuck this! Your command is not well-formed.\n" "Check my profile description or " "[fuck around with him...](/posts/0f99d95040130133bbca14dae9b624ce)") # mark as read if it # is not a mention n.mark() client.logout()
#! /usr/bin/env python from time import gmtime, strftime from foaas import foaas from diaspy_client import Client import re import urllib2 client = Client() notify = client.notifications() for n in notify: if not n.unread: continue idm = re.search('href=\\"/posts/(\d+?)\\"', n._data['note_html']) if hasattr(idm, 'group'): post_id = idm.group(1) text = client.getPostText(post_id) m = re.search('^\s*\@\{[^\};]+;[^\};]+\}\s+(\/.+?)$', text) if hasattr(m, 'group'): try: command = m.group(1) client.post(foaas(command)) except urllib2.URLError: client.comment(post_id, "Fuck this! Something went wrong :\\") else: client.comment(post_id, "Fuck this! Your command is not well-formed.\n" "Check my [profile description](/people/448c48d02c1c013349f314dae9b624ce) or " "[fuck around with him...](/posts/0f99d95040130133bbca14dae9b624ce)") # mark as read if it # is not a mention n.mark() client.logout()
Add link to foaas profile
Add link to foaas profile
Python
mit
Zauberstuhl/foaasBot
--- +++ @@ -26,7 +26,7 @@ else: client.comment(post_id, "Fuck this! Your command is not well-formed.\n" - "Check my profile description or " + "Check my [profile description](/people/448c48d02c1c013349f314dae9b624ce) or " "[fuck around with him...](/posts/0f99d95040130133bbca14dae9b624ce)") # mark as read if it
2c4e93844c1b704e3435816737a6cfda624ff7b7
bot.py
bot.py
import tornado.httpserver import tornado.ioloop import tornado.web from tornado.options import define, options from settings import * class MainHandler(tornado.web.RequestHandler): def get(self): self.write("Hello, world") application = tornado.web.Application([ (r'/', MainHandler), ]) if __name__ == "__main__": define("port", default="443", help="Port to listen on") define("host", default="localhost", help="Server address to listen on") tornado.options.parse_command_line() http_server = tornado.httpserver.HTTPServer(application, ssl_options={ "certfile": PEMFILE, "keyfile": KEYFILE }) http_server.listen(int(options.port), address=options.host) tornado.ioloop.IOLoop.current().start()
import json import logging import tornado.httpserver import tornado.ioloop import tornado.web from tornado.options import define, options from settings import * class MainHandler(tornado.web.RequestHandler): def post(self): logging.debug(json.dumps(self.request)) application = tornado.web.Application([ (r'/', MainHandler), ]) if __name__ == "__main__": define("port", default="443", help="Port to listen on") define("host", default="localhost", help="Server address to listen on") tornado.options.parse_command_line() http_server = tornado.httpserver.HTTPServer(application, ssl_options={ "certfile": PEMFILE, "keyfile": KEYFILE }) http_server.listen(int(options.port), address=options.host) tornado.ioloop.IOLoop.current().start()
Debug any incoming POST request.
Debug any incoming POST request.
Python
mit
pistonsky/pistonskybot
--- +++ @@ -1,3 +1,5 @@ +import json +import logging import tornado.httpserver import tornado.ioloop import tornado.web @@ -8,8 +10,8 @@ class MainHandler(tornado.web.RequestHandler): - def get(self): - self.write("Hello, world") + def post(self): + logging.debug(json.dumps(self.request)) application = tornado.web.Application([ (r'/', MainHandler),
ea4949dab887a14a0ca8f5ffbcd3c578c61c005e
api/urls.py
api/urls.py
from django.conf.urls import include, url from api import views urlpatterns = [ url(r'^services/$', views.services, name='api-services'), url(r'^collect/$', views.collect_response, name='api-collect'), url(r'^search/text/$', views.text_search, name='api-text-search'), url(r'^license/$', views.licensing, name='api-licensing'), # Oauth2 urls url(r'^o/', include('api.oauth2_urls', namespace='oauth2_provider')), ]
from django.conf.urls import include, url from api import views urlpatterns = [ url(r'^v1/services/$', views.services, name='api-services'), url(r'^v1/collect/$', views.collect_response, name='api-collect'), url(r'^v1/search/text/$', views.text_search, name='api-text-search'), url(r'^v1/license/$', views.licensing, name='api-licensing'), # Oauth2 urls url(r'^o/', include('api.oauth2_urls', namespace='oauth2_provider')), ]
Add API versioning at the url
Add API versioning at the url https://github.com/AudioCommons/ac-mediator/issues/13
Python
apache-2.0
AudioCommons/ac-mediator,AudioCommons/ac-mediator,AudioCommons/ac-mediator
--- +++ @@ -2,10 +2,10 @@ from api import views urlpatterns = [ - url(r'^services/$', views.services, name='api-services'), - url(r'^collect/$', views.collect_response, name='api-collect'), - url(r'^search/text/$', views.text_search, name='api-text-search'), - url(r'^license/$', views.licensing, name='api-licensing'), + url(r'^v1/services/$', views.services, name='api-services'), + url(r'^v1/collect/$', views.collect_response, name='api-collect'), + url(r'^v1/search/text/$', views.text_search, name='api-text-search'), + url(r'^v1/license/$', views.licensing, name='api-licensing'), # Oauth2 urls url(r'^o/', include('api.oauth2_urls', namespace='oauth2_provider')),
21193559b063e85f26971d5ae6181a0bd097cda3
tests/utilities_test.py
tests/utilities_test.py
#pylint: disable=W0104,W0108 import pytest import pyop import numpy as np ####################################################################### # Tests # ####################################################################### def testEnsure2dColumn(capsys): @pyop.ensure2dColumn def printShape(x): print(x.shape) return x input_vec = np.random.rand(10) output = printShape(input_vec) print_out, _ = capsys.readouterr() np.testing.assert_allclose(input_vec, output) assert print_out == "(10, 1)\n" input_vec = np.random.rand(10, 10) output = printShape(input_vec) print_out, _ = capsys.readouterr() np.testing.assert_allclose(input_vec, output) assert print_out == "(10, 10)\n"
#pylint: disable=W0104,W0108 import pyop import numpy as np ####################################################################### # Tests # ####################################################################### def testEnsure2dColumn(capsys): @pyop.ensure2dColumn def printShape(x): print(x.shape) return x input_vec = np.random.rand(10) output = printShape(input_vec) print_out, _ = capsys.readouterr() np.testing.assert_allclose(input_vec, output) assert print_out == "(10, 1)\n" input_vec = np.random.rand(10, 10) output = printShape(input_vec) print_out, _ = capsys.readouterr() np.testing.assert_allclose(input_vec, output) assert print_out == "(10, 10)\n" ############ # Vector # ############ @pyop.vector def multFirstColumn(column): img = column.reshape((2, 2), order = 'C') img[:, 0] *= 2 return img.flatten(0) def testVectorOnMatrix(): np.testing.assert_allclose( multFirstColumn(np.array([[1, 1, 1, 1], [2, 1, 2, 1]]).T), np.array([[2, 4], [1, 1], [2, 4], [1, 1]])) def testVectorOnVector(): np.testing.assert_allclose( multFirstColumn(np.array([1, 1, 1, 1])), np.array(np.array([2, 1, 2, 1])))
Test vector, passes matrix and vector input.
Test vector, passes matrix and vector input.
Python
bsd-3-clause
ryanorendorff/pyop
--- +++ @@ -1,5 +1,4 @@ #pylint: disable=W0104,W0108 -import pytest import pyop import numpy as np @@ -29,3 +28,25 @@ np.testing.assert_allclose(input_vec, output) assert print_out == "(10, 10)\n" + + +############ +# Vector # +############ +@pyop.vector +def multFirstColumn(column): + img = column.reshape((2, 2), order = 'C') + img[:, 0] *= 2 + return img.flatten(0) + + +def testVectorOnMatrix(): + np.testing.assert_allclose( + multFirstColumn(np.array([[1, 1, 1, 1], [2, 1, 2, 1]]).T), + np.array([[2, 4], [1, 1], [2, 4], [1, 1]])) + + +def testVectorOnVector(): + np.testing.assert_allclose( + multFirstColumn(np.array([1, 1, 1, 1])), + np.array(np.array([2, 1, 2, 1])))
98a6b04b83843861003862d835b3a2f6f2364506
tests/context_tests.py
tests/context_tests.py
# -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function, unicode_literals from mock import Mock from nose.tools import eq_ from cg.context import Context, ContextFactory class TestContextFactory(object): def test_context_gets_created_correctly(self): handle = 123 bridge = Mock() bridge.cgCreateContext.return_value = handle cf = ContextFactory(bridge) context = cf.create() eq_(context._cgcontext, handle) def test_opengl_states_are_set_on_context_creation(self): handle = 'x' bridge = Mock() bridge.cgCreateContext.return_value = handle cf = ContextFactory(bridge) context = cf.create() bridge.cgGLRegisterStates.assert_called_once_with(handle) class TestContext(object): def test_context_disposes_of_its_resources(self): handle = 234 bridge = Mock() context = Context(handle, bridge) context.dispose() bridge.cgDestroyContext.assert_called_once_with(handle)
# -*- coding: utf-8 -*- from __future__ import absolute_import, division, print_function, unicode_literals from mock import Mock from nose.tools import eq_ from cg.context import Context, ContextFactory class TestContextFactory(object): def test_context_gets_created_correctly(self): handle = 123 bridge = Mock() bridge.cgCreateContext.return_value = handle cf = ContextFactory(bridge) context = cf.create() eq_(context._cgcontext, handle) class TestContext(object): def test_context_disposes_of_its_resources(self): handle = 234 bridge = Mock() context = Context(handle, bridge) context.dispose() bridge.cgDestroyContext.assert_called_once_with(handle)
Update test to match last changes in the API
Update test to match last changes in the API
Python
mit
jstasiak/python-cg,jstasiak/python-cg
--- +++ @@ -16,16 +16,6 @@ context = cf.create() eq_(context._cgcontext, handle) - def test_opengl_states_are_set_on_context_creation(self): - handle = 'x' - bridge = Mock() - bridge.cgCreateContext.return_value = handle - - cf = ContextFactory(bridge) - context = cf.create() - - bridge.cgGLRegisterStates.assert_called_once_with(handle) - class TestContext(object): def test_context_disposes_of_its_resources(self): handle = 234
4fb1023c461498a080d371e50a5a4971924cb1bc
third_party/__init__.py
third_party/__init__.py
import os.path import sys # This bit of evil should inject third_party into the path for relative imports. sys.path.append(os.path.dirname(__file__))
import os.path import sys # This bit of evil should inject third_party into the path for relative imports. sys.path.insert(1, os.path.dirname(__file__))
Insert third_party into the second slot of sys.path rather than the last slot
Insert third_party into the second slot of sys.path rather than the last slot
Python
apache-2.0
somehume/namebench
--- +++ @@ -2,4 +2,4 @@ import sys # This bit of evil should inject third_party into the path for relative imports. -sys.path.append(os.path.dirname(__file__)) +sys.path.insert(1, os.path.dirname(__file__))
377beee13a8cd0ca23f8f2e37dd2816571721921
tests/sources_tests.py
tests/sources_tests.py
import os import subprocess from nose.tools import istest, assert_equal from whack.sources import PackageSourceFetcher from whack.tempdir import create_temporary_dir from whack.files import read_file, write_file @istest def can_fetch_package_source_from_source_control(): with create_temporary_dir() as package_source_dir: write_file(os.path.join(package_source_dir, "name"), "Bob") _convert_to_git_repo(package_source_dir) source_fetcher = PackageSourceFetcher([]) repo_uri = "git+file://{0}".format(package_source_dir) with source_fetcher.fetch(repo_uri) as package_source: assert_equal("Bob", read_file(os.path.join(package_source.path, "name"))) def _convert_to_git_repo(cwd): def _git(command): subprocess.check_call(["git"] + command, cwd=cwd) _git(["init"]) _git(["add", "."]) _git(["commit", "-m", "Initial commit"])
import os import subprocess from nose.tools import istest, assert_equal from whack.sources import PackageSourceFetcher from whack.tempdir import create_temporary_dir from whack.files import read_file, write_file @istest def can_fetch_package_source_from_source_control(): with create_temporary_dir() as package_source_dir: write_file(os.path.join(package_source_dir, "name"), "Bob") _convert_to_git_repo(package_source_dir) source_fetcher = PackageSourceFetcher([]) repo_uri = "git+file://{0}".format(package_source_dir) with source_fetcher.fetch(repo_uri) as package_source: assert_equal("Bob", read_file(os.path.join(package_source.path, "name"))) @istest def can_fetch_package_source_from_local_path(): with create_temporary_dir() as package_source_dir: write_file(os.path.join(package_source_dir, "name"), "Bob") source_fetcher = PackageSourceFetcher([]) with source_fetcher.fetch(package_source_dir) as package_source: assert_equal("Bob", read_file(os.path.join(package_source.path, "name"))) def _convert_to_git_repo(cwd): def _git(command): subprocess.check_call(["git"] + command, cwd=cwd) _git(["init"]) _git(["add", "."]) _git(["commit", "-m", "Initial commit"])
Add test for fetching local package sources
Add test for fetching local package sources
Python
bsd-2-clause
mwilliamson/whack
--- +++ @@ -18,6 +18,16 @@ repo_uri = "git+file://{0}".format(package_source_dir) with source_fetcher.fetch(repo_uri) as package_source: assert_equal("Bob", read_file(os.path.join(package_source.path, "name"))) + + +@istest +def can_fetch_package_source_from_local_path(): + with create_temporary_dir() as package_source_dir: + write_file(os.path.join(package_source_dir, "name"), "Bob") + + source_fetcher = PackageSourceFetcher([]) + with source_fetcher.fetch(package_source_dir) as package_source: + assert_equal("Bob", read_file(os.path.join(package_source.path, "name")))
c4db09cd2d4dac37afcf75d5cf4d8c8f881aac2d
tests/test_examples.py
tests/test_examples.py
'''Search all our doc comments for "Example" blocks and try executing them.''' import re import sourcer def run_examples(package): pattern = re.compile(r''' (\s*) # initial indent Example # magic keyword ([^\n]*) # optional description \:\: # magic marker # Each line of the example is indented # by four additional spaces: (\n((\1\ \ \ \ .*)?\n)+) ''', re.IGNORECASE | re.VERBOSE ) for k, v in package.__dict__.iteritems(): if k.startswith('__') and k.endswith('__'): continue doc = getattr(v, '__doc__') or '' for m in pattern.finditer(doc): indent = '\n ' + m.group(1) body = m.group(3) example = body.replace(indent, '\n') print ' Running', k, 'example', m.group(2).strip() exec example in {} if __name__ == '__main__': run_examples(sourcer)
'''Search all our doc comments for "Example" blocks and try executing them.''' import re import sourcer.expressions def run_examples(package): pattern = re.compile(r''' (\s*) # initial indent Example # magic keyword ([^\n]*) # optional description \:\: # magic marker # Each line of the example is indented # by four additional spaces: (\n((\1\ \ \ \ .*)?\n)+) ''', re.IGNORECASE | re.VERBOSE ) for k, v in package.__dict__.iteritems(): if k.startswith('__') and k.endswith('__'): continue doc = getattr(v, '__doc__') or '' for m in pattern.finditer(doc): indent = '\n ' + m.group(1) body = m.group(3) example = body.replace(indent, '\n') print ' Running', k, 'example', m.group(2).strip() exec example in {} if __name__ == '__main__': run_examples(sourcer.expressions)
Test the doc comments in the expressions module.
Test the doc comments in the expressions module.
Python
mit
jvs/sourcer
--- +++ @@ -1,6 +1,6 @@ '''Search all our doc comments for "Example" blocks and try executing them.''' import re -import sourcer +import sourcer.expressions def run_examples(package): @@ -30,4 +30,4 @@ if __name__ == '__main__': - run_examples(sourcer) + run_examples(sourcer.expressions)
aaae301f62b4e0b3cdd5d1756a03b619a8f18222
tests/test_hamilton.py
tests/test_hamilton.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """Tests for the Hamiltonian method.""" import pytest from numpy.testing import assert_allclose from parameters import KPT, T_VALUES @pytest.mark.parametrize("kpt", KPT) @pytest.mark.parametrize("t_values", T_VALUES) @pytest.mark.parametrize("convention", [1, 2]) def test_simple_hamilton(get_model, kpt, t_values, compare_isclose, convention): """ Regression test for the Hamiltonian of a simple model. """ model = get_model(*t_values) compare_isclose(model.hamilton(kpt, convention=convention)) @pytest.mark.parametrize("t_values", T_VALUES) @pytest.mark.parametrize("convention", [1, 2]) def test_parallel_hamilton(get_model, t_values, convention): """ Test that passing multiple k-points to the Hamiltonian gives the same results as evaluating them individually. """ model = get_model(*t_values) assert_allclose( model.hamilton(KPT, convention=convention), [model.hamilton(k, convention=convention) for k in KPT], )
#!/usr/bin/env python # -*- coding: utf-8 -*- """Tests for the Hamiltonian method.""" import pytest from numpy.testing import assert_allclose from parameters import KPT, T_VALUES @pytest.mark.parametrize("kpt", KPT) @pytest.mark.parametrize("t_values", T_VALUES) @pytest.mark.parametrize("convention", [1, 2]) def test_simple_hamilton(get_model, kpt, t_values, compare_isclose, convention): """ Regression test for the Hamiltonian of a simple model. """ model = get_model(*t_values) compare_isclose(model.hamilton(kpt, convention=convention)) @pytest.mark.parametrize("t_values", T_VALUES) @pytest.mark.parametrize("convention", [1, 2]) def test_parallel_hamilton(get_model, t_values, convention): """ Test that passing multiple k-points to the Hamiltonian gives the same results as evaluating them individually. """ model = get_model(*t_values) assert_allclose( model.hamilton(KPT, convention=convention), [model.hamilton(k, convention=convention) for k in KPT], ) @pytest.mark.parametrize("convention", ["a", "1", None]) def test_invalid_convention(get_model, convention): """ Test that giving an invalid 'convention' raises an error. """ model = get_model(t1=0, t2=0.1) with pytest.raises(ValueError): model.hamilton((0, 0, 0), convention=convention)
Add test for invalid 'convention' in hamilton method
Add test for invalid 'convention' in hamilton method
Python
apache-2.0
Z2PackDev/TBmodels,Z2PackDev/TBmodels
--- +++ @@ -31,3 +31,13 @@ model.hamilton(KPT, convention=convention), [model.hamilton(k, convention=convention) for k in KPT], ) + + +@pytest.mark.parametrize("convention", ["a", "1", None]) +def test_invalid_convention(get_model, convention): + """ + Test that giving an invalid 'convention' raises an error. + """ + model = get_model(t1=0, t2=0.1) + with pytest.raises(ValueError): + model.hamilton((0, 0, 0), convention=convention)
08730308134d0a15be996e8e7bb1a19bc1930f12
tests/test_plotting.py
tests/test_plotting.py
from contextlib import contextmanager import os import tempfile import unittest from matplotlib.pyplot import Artist, savefig from shapely.geometry import Polygon, LineString, Point from geopandas import GeoSeries @contextmanager def get_tempfile(): f, path = tempfile.mkstemp() try: yield path finally: try: os.remove(path) except: pass class TestSeriesPlot(unittest.TestCase): def setUp(self): self.t1 = Polygon([(0, 0), (1, 0), (1, 1)]) self.t2 = Polygon([(1, 0), (2, 1), (2, 1)]) self.polys = GeoSeries([self.t1, self.t2]) def test_poly_plot(self): """ Test plotting a simple series of polygons """ ax = self.polys.plot() self.assertIsInstance(ax, Artist) with get_tempfile() as file: savefig(file) if __name__ == '__main__': unittest.main()
import os import unittest from matplotlib.pyplot import Artist, savefig from matplotlib.testing.decorators import image_comparison from shapely.geometry import Polygon, LineString, Point from geopandas import GeoSeries # If set to True, generate images rather than perform tests (all tests will pass!) GENERATE_BASELINE = False BASELINE_DIR = os.path.join(os.path.dirname(__file__), 'baseline_images', 'test_plotting') def save_baseline_image(filename): """ save a baseline image """ savefig(os.path.join(BASELINE_DIR, filename)) @image_comparison(baseline_images=['poly_plot'], extensions=['png']) def test_poly_plot(): """ Test plotting a simple series of polygons """ t1 = Polygon([(0, 0), (1, 0), (1, 1)]) t2 = Polygon([(1, 0), (2, 1), (2, 1)]) polys = GeoSeries([t1, t2]) ax = polys.plot() assert isinstance(ax, Artist) if GENERATE_BASELINE: save_baseline_image('poly_plot.png') if __name__ == '__main__': import nose nose.runmodule(argv=['-s', '--with-doctest'])
Use matplotlib image_comparison to actually compare plot output
TST: Use matplotlib image_comparison to actually compare plot output
Python
bsd-3-clause
jorisvandenbossche/geopandas,jorisvandenbossche/geopandas,jdmcbr/geopandas,koldunovn/geopandas,ozak/geopandas,maxalbert/geopandas,IamJeffG/geopandas,ozak/geopandas,geopandas/geopandas,fonnesbeck/geopandas,kwinkunks/geopandas,geopandas/geopandas,geopandas/geopandas,jorisvandenbossche/geopandas,urschrei/geopandas,snario/geopandas,scw/geopandas,jdmcbr/geopandas,micahcochran/geopandas,micahcochran/geopandas,perrygeo/geopandas
--- +++ @@ -1,38 +1,32 @@ -from contextlib import contextmanager import os -import tempfile import unittest from matplotlib.pyplot import Artist, savefig +from matplotlib.testing.decorators import image_comparison from shapely.geometry import Polygon, LineString, Point from geopandas import GeoSeries +# If set to True, generate images rather than perform tests (all tests will pass!) +GENERATE_BASELINE = False -@contextmanager -def get_tempfile(): - f, path = tempfile.mkstemp() - try: - yield path - finally: - try: - os.remove(path) - except: - pass +BASELINE_DIR = os.path.join(os.path.dirname(__file__), 'baseline_images', 'test_plotting') -class TestSeriesPlot(unittest.TestCase): +def save_baseline_image(filename): + """ save a baseline image """ + savefig(os.path.join(BASELINE_DIR, filename)) - def setUp(self): - self.t1 = Polygon([(0, 0), (1, 0), (1, 1)]) - self.t2 = Polygon([(1, 0), (2, 1), (2, 1)]) - self.polys = GeoSeries([self.t1, self.t2]) - - def test_poly_plot(self): - """ Test plotting a simple series of polygons """ - ax = self.polys.plot() - self.assertIsInstance(ax, Artist) - with get_tempfile() as file: - savefig(file) +@image_comparison(baseline_images=['poly_plot'], extensions=['png']) +def test_poly_plot(): + """ Test plotting a simple series of polygons """ + t1 = Polygon([(0, 0), (1, 0), (1, 1)]) + t2 = Polygon([(1, 0), (2, 1), (2, 1)]) + polys = GeoSeries([t1, t2]) + ax = polys.plot() + assert isinstance(ax, Artist) + if GENERATE_BASELINE: + save_baseline_image('poly_plot.png') if __name__ == '__main__': - unittest.main() + import nose + nose.runmodule(argv=['-s', '--with-doctest'])
331060f37841dccbd8974f01c063dc1b5c112121
formula/openssl.py
formula/openssl.py
import winbrew class Openssl(winbrew.Formula): url = 'http://www.openssl.org/source/openssl-1.0.1f.tar.gz' homepage = 'http://www.openssl.org' sha1 = '' build_deps = () deps = () def install(self): self.system('perl Configure VC-WIN32 no-asm --prefix=C:\\Winbrew\\lib\\OpenSSL') self.system('ms\\\\do_ms.bat') self.system('nmake -f ms\\\\nt.mak') self.lib('out32\\libeay32.lib') self.lib('out32\\ssleay32.lib') self.includes('include\\openssl', dest='openssl') #self.system('nmake -f ms\\ntdll.mak install') # Dynamic libraries def test(self): self.system('nmake -f ms\\\\nt.mak test')
import winbrew class Openssl(winbrew.Formula): url = 'http://www.openssl.org/source/openssl-1.0.1g.tar.gz' homepage = 'http://www.openssl.org' sha1 = '' build_deps = () deps = () def install(self): self.system('perl Configure VC-WIN32 no-asm --prefix=C:\\Winbrew\\lib\\OpenSSL') self.system('ms\\\\do_ms.bat') self.system('nmake -f ms\\\\nt.mak') self.lib('out32\\libeay32.lib') self.lib('out32\\ssleay32.lib') self.includes('include\\openssl', dest='openssl') #self.system('nmake -f ms\\ntdll.mak install') # Dynamic libraries def test(self): self.system('nmake -f ms\\\\nt.mak test')
Upgrade to OpenSSL 1.0.1g to avoid heartbleed bug
Upgrade to OpenSSL 1.0.1g to avoid heartbleed bug
Python
mit
mfichman/winbrew
--- +++ @@ -1,7 +1,7 @@ import winbrew class Openssl(winbrew.Formula): - url = 'http://www.openssl.org/source/openssl-1.0.1f.tar.gz' + url = 'http://www.openssl.org/source/openssl-1.0.1g.tar.gz' homepage = 'http://www.openssl.org' sha1 = '' build_deps = ()
5b8b210a73282f6176883f3fab1dd0b2801b3f34
wsgi/app.py
wsgi/app.py
# flake8: noqa # newrelic import & initialization must come first # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration try: import newrelic.agent except ImportError: newrelic = False if newrelic: newrelic_ini = config('NEWRELIC_PYTHON_INI_FILE', default='') if newrelic_ini: newrelic.agent.initialize(newrelic_ini) else: newrelic = False import os from bedrock.base.config_manager import config IS_HTTPS = os.environ.get('HTTPS', '').strip() == 'on' os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'bedrock.settings') # must be imported after env var is set above. from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from whitenoise.django import DjangoWhiteNoise from raven.contrib.django.raven_compat.middleware.wsgi import Sentry class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest application = DjangoWhiteNoise(application) application = Sentry(application) if newrelic: application = newrelic.agent.wsgi_application()(application)
# flake8: noqa # newrelic import & initialization must come first # https://docs.newrelic.com/docs/agents/python-agent/installation/python-agent-advanced-integration#manual-integration try: import newrelic.agent except ImportError: newrelic = False else: newrelic.agent.initialize() import os from bedrock.base.config_manager import config IS_HTTPS = os.environ.get('HTTPS', '').strip() == 'on' os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'bedrock.settings') # must be imported after env var is set above. from django.core.handlers.wsgi import WSGIRequest from django.core.wsgi import get_wsgi_application from whitenoise.django import DjangoWhiteNoise from raven.contrib.django.raven_compat.middleware.wsgi import Sentry class WSGIHTTPSRequest(WSGIRequest): def _get_scheme(self): if IS_HTTPS: return 'https' return super(WSGIHTTPSRequest, self)._get_scheme() application = get_wsgi_application() application.request_class = WSGIHTTPSRequest application = DjangoWhiteNoise(application) application = Sentry(application) if newrelic: application = newrelic.agent.wsgi_application()(application)
Remove unused ability to use custom newrelic.ini
Remove unused ability to use custom newrelic.ini
Python
mpl-2.0
flodolo/bedrock,craigcook/bedrock,hoosteeno/bedrock,sylvestre/bedrock,craigcook/bedrock,pascalchevrel/bedrock,kyoshino/bedrock,kyoshino/bedrock,sgarrity/bedrock,ericawright/bedrock,ericawright/bedrock,MichaelKohler/bedrock,mozilla/bedrock,alexgibson/bedrock,alexgibson/bedrock,hoosteeno/bedrock,ericawright/bedrock,sgarrity/bedrock,flodolo/bedrock,flodolo/bedrock,mozilla/bedrock,pascalchevrel/bedrock,sgarrity/bedrock,craigcook/bedrock,sylvestre/bedrock,MichaelKohler/bedrock,hoosteeno/bedrock,kyoshino/bedrock,sylvestre/bedrock,MichaelKohler/bedrock,alexgibson/bedrock,ericawright/bedrock,craigcook/bedrock,mozilla/bedrock,alexgibson/bedrock,kyoshino/bedrock,sgarrity/bedrock,flodolo/bedrock,MichaelKohler/bedrock,hoosteeno/bedrock,mozilla/bedrock,sylvestre/bedrock,pascalchevrel/bedrock,pascalchevrel/bedrock
--- +++ @@ -5,14 +5,9 @@ import newrelic.agent except ImportError: newrelic = False +else: + newrelic.agent.initialize() - -if newrelic: - newrelic_ini = config('NEWRELIC_PYTHON_INI_FILE', default='') - if newrelic_ini: - newrelic.agent.initialize(newrelic_ini) - else: - newrelic = False import os
0e533ad0cc42431a57758b577cf96783ee4b7484
spacy/tests/test_download.py
spacy/tests/test_download.py
# coding: utf-8 from __future__ import unicode_literals from ..download import download, get_compatibility, get_version, check_error_depr import pytest def test_download_fetch_compatibility(): compatibility = get_compatibility() assert type(compatibility) == dict @pytest.mark.slow @pytest.mark.parametrize('model', ['en_core_web_md-1.2.0']) def test_download_direct_download(model): download(model, direct=True) @pytest.mark.parametrize('model', ['en_core_web_md']) def test_download_get_matching_version_succeeds(model): comp = { model: ['1.7.0', '0.100.0'] } assert get_version(model, comp) @pytest.mark.parametrize('model', ['en_core_web_md']) def test_download_get_matching_version_fails(model): diff_model = 'test_' + model comp = { diff_model: ['1.7.0', '0.100.0'] } with pytest.raises(SystemExit): assert get_version(model, comp) @pytest.mark.parametrize('model', [False, None, '', 'all']) def test_download_no_model_depr_error(model): with pytest.raises(SystemExit): check_error_depr(model)
# coding: utf-8 from __future__ import unicode_literals from ..download import download, get_compatibility, get_version, check_error_depr import pytest @pytest.mark.slow def test_download_fetch_compatibility(): compatibility = get_compatibility() assert type(compatibility) == dict @pytest.mark.slow @pytest.mark.parametrize('model', ['en_core_web_md-1.2.0']) def test_download_direct_download(model): download(model, direct=True) @pytest.mark.parametrize('model', ['en_core_web_md']) def test_download_get_matching_version_succeeds(model): comp = { model: ['1.7.0', '0.100.0'] } assert get_version(model, comp) @pytest.mark.parametrize('model', ['en_core_web_md']) def test_download_get_matching_version_fails(model): diff_model = 'test_' + model comp = { diff_model: ['1.7.0', '0.100.0'] } with pytest.raises(SystemExit): assert get_version(model, comp) @pytest.mark.parametrize('model', [False, None, '', 'all']) def test_download_no_model_depr_error(model): with pytest.raises(SystemExit): check_error_depr(model)
Mark compatibility table test as slow (temporary)
Mark compatibility table test as slow (temporary) Prevent Travis from running test test until models repo is published
Python
mit
oroszgy/spaCy.hu,oroszgy/spaCy.hu,aikramer2/spaCy,spacy-io/spaCy,raphael0202/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,aikramer2/spaCy,spacy-io/spaCy,explosion/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,spacy-io/spaCy,spacy-io/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,raphael0202/spaCy,recognai/spaCy,honnibal/spaCy,explosion/spaCy,aikramer2/spaCy,raphael0202/spaCy,explosion/spaCy,recognai/spaCy,aikramer2/spaCy,explosion/spaCy,spacy-io/spaCy,recognai/spaCy,recognai/spaCy,honnibal/spaCy,aikramer2/spaCy,spacy-io/spaCy,honnibal/spaCy,Gregory-Howard/spaCy,raphael0202/spaCy,raphael0202/spaCy,explosion/spaCy,honnibal/spaCy,explosion/spaCy,recognai/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,Gregory-Howard/spaCy,Gregory-Howard/spaCy
--- +++ @@ -5,6 +5,7 @@ import pytest +@pytest.mark.slow def test_download_fetch_compatibility(): compatibility = get_compatibility() assert type(compatibility) == dict
a6a4c2920abd099a97839584b96af10dcd25afe2
tests/test_public_api.py
tests/test_public_api.py
# This file is part of python-markups test suite # License: BSD # Copyright: (C) Dmitry Shachnev, 2012-2015 import markups import unittest class APITest(unittest.TestCase): def test_api(self): all_markups = markups.get_all_markups() self.assertIn(markups.MarkdownMarkup, all_markups) self.assertIn(markups.ReStructuredTextMarkup, all_markups) markup_class = markups.find_markup_class_by_name('restructuredtext') self.assertEqual(markups.ReStructuredTextMarkup, markup_class) markup_class = markups.get_markup_for_file_name('myfile.mkd', return_class=True) self.assertEqual(markups.MarkdownMarkup, markup_class) @unittest.skipUnless(markups.MarkdownMarkup.available(), 'Markdown not available') def test_api_instance(self): markup = markups.get_markup_for_file_name('myfile.mkd') self.assertIsInstance(markup, markups.MarkdownMarkup) if __name__ == '__main__': unittest.main()
# This file is part of python-markups test suite # License: BSD # Copyright: (C) Dmitry Shachnev, 2012-2015 import markups import unittest class APITest(unittest.TestCase): def test_api(self): all_markups = markups.get_all_markups() self.assertIn(markups.MarkdownMarkup, all_markups) self.assertIn(markups.ReStructuredTextMarkup, all_markups) markup_class = markups.find_markup_class_by_name('restructuredtext') self.assertEqual(markups.ReStructuredTextMarkup, markup_class) markup_class = markups.get_markup_for_file_name('myfile.mkd', return_class=True) self.assertEqual(markups.MarkdownMarkup, markup_class) @unittest.skipUnless(markups.MarkdownMarkup.available(), 'Markdown not available') def test_api_instance(self): markup = markups.get_markup_for_file_name('myfile.mkd') self.assertIsInstance(markup, markups.MarkdownMarkup) @unittest.skipUnless(markups.MarkdownMarkup.available(), 'Markdown not available') def test_available_markups(self): available_markups = markups.get_available_markups() self.assertIn(markups.MarkdownMarkup, available_markups) if __name__ == '__main__': unittest.main()
Add a test for get_available_markups() function
Add a test for get_available_markups() function
Python
bsd-3-clause
mitya57/pymarkups,retext-project/pymarkups
--- +++ @@ -20,5 +20,10 @@ markup = markups.get_markup_for_file_name('myfile.mkd') self.assertIsInstance(markup, markups.MarkdownMarkup) + @unittest.skipUnless(markups.MarkdownMarkup.available(), 'Markdown not available') + def test_available_markups(self): + available_markups = markups.get_available_markups() + self.assertIn(markups.MarkdownMarkup, available_markups) + if __name__ == '__main__': unittest.main()
9ce5a020ac6e9bbdf7e2fc0c34c98cdfaf9e0a45
tests/formatters/conftest.py
tests/formatters/conftest.py
import npc import pytest @pytest.fixture(scope="module") def character(): char = npc.character.Character() char.append('description', 'Fee fie foe fum') char.append('type', 'human') return char
import npc import pytest @pytest.fixture(scope="module") def character(): char = npc.character.Character() char.tags('description').append('Fee fie foe fum') char.tags('type').append('human') return char
Set up defaults using tag syntax
Set up defaults using tag syntax
Python
mit
aurule/npc,aurule/npc
--- +++ @@ -4,6 +4,6 @@ @pytest.fixture(scope="module") def character(): char = npc.character.Character() - char.append('description', 'Fee fie foe fum') - char.append('type', 'human') + char.tags('description').append('Fee fie foe fum') + char.tags('type').append('human') return char
69ec55e0a6b4d314eb1381afc09236a5aa01d3d8
util/git-author-comm.py
util/git-author-comm.py
#!/usr/bin/env python # Display authors which appear as contributors in both (two) repositories. import os,sys def usage(): print 'Show authors which appear in two git repositories.' print 'python2 git-author-comm.py [path-to-git-repo] [path-to-git-repo]' def sysout(command): return os.popen(command).read() def get_authors(git): cmd = "git --work-tree="+git+" --git-dir="+git+"/.git log --format='%aN' | sort -u" # Split results to a list. authors = sysout(cmd).split('\n') # Remove any empty lines prior to return. return filter(None, authors) def get_same_authors(a1,a2): # Return a set of authors which appear in both repositories. return set(a1).intersection(a2) if __name__ == "__main__": if len(sys.argv) != 3: usage() else: git1 = sys.argv[1] git2 = sys.argv[2] authors = get_same_authors(get_authors(git1),get_authors(git2)) if len(authors) == 0: print 'There are no matches.' else: print 'The folowing appear in both repositories: ' for author in authors: print author
#!/usr/bin/env python # Display authors which appear as contributors in both (two) repositories. import os,sys def usage(): print 'Show authors which appear in two local git repositories.' print 'python2 git-author-comm.py [path-to-local-git-repo] [path-to-local-git-repo]' def sysout(command): return os.popen(command).read() def get_authors(git): cmd = "git --work-tree="+git+" --git-dir="+git+"/.git log --format='%aN' | sort -u" # Split results to a list. authors = sysout(cmd).split('\n') # Remove any empty lines prior to return. return filter(None, authors) def get_same_authors(a1,a2): # Return a set of authors which appear in both repositories. return set(a1).intersection(a2) if __name__ == "__main__": if len(sys.argv) != 3: usage() else: git1 = sys.argv[1] git2 = sys.argv[2] authors = get_same_authors(get_authors(git1),get_authors(git2)) if len(authors) == 0: print 'There are no matches.' else: print 'The folowing appear in both repositories: ' for author in authors: print author
Make it more obvious that you need to use a local repository as an argument, not a url
Make it more obvious that you need to use a local repository as an argument, not a url
Python
mit
baykovr/toolbox,baykovr/toolbox,baykovr/toolbox,baykovr/toolbox,baykovr/toolbox
--- +++ @@ -6,8 +6,8 @@ import os,sys def usage(): - print 'Show authors which appear in two git repositories.' - print 'python2 git-author-comm.py [path-to-git-repo] [path-to-git-repo]' + print 'Show authors which appear in two local git repositories.' + print 'python2 git-author-comm.py [path-to-local-git-repo] [path-to-local-git-repo]' def sysout(command): return os.popen(command).read()
4669a033ee4fbde5e3c2447778657a20a73d5df8
thefuck/shells/powershell.py
thefuck/shells/powershell.py
from .generic import Generic class Powershell(Generic): def app_alias(self, fuck): return 'function ' + fuck + ' { \n' \ ' $fuck = $(thefuck (Get-History -Count 1).CommandLine);\n' \ ' if (-not [string]::IsNullOrWhiteSpace($fuck)) {\n' \ ' if ($fuck.StartsWith("echo")) { $fuck = $fuck.Substring(5); }\n' \ ' else { iex "$fuck"; }\n' \ ' }\n' \ '}\n' def and_(self, *commands): return u' -and '.join('({0})'.format(c) for c in commands) def how_to_configure(self): return { 'content': 'iex "thefuck --alias"', 'path': '$profile', 'reload': '& $profile', }
from .generic import Generic class Powershell(Generic): def app_alias(self, fuck): return 'function ' + fuck + ' {\n' \ ' $history = (Get-History -Count 1).CommandLine;\n' \ ' if (-not [string]::IsNullOrWhiteSpace($history)) {\n' \ ' $fuck = $(thefuck $history);\n' \ ' if (-not [string]::IsNullOrWhiteSpace($fuck)) {\n' \ ' if ($fuck.StartsWith("echo")) { $fuck = $fuck.Substring(5); }\n' \ ' else { iex "$fuck"; }\n' \ ' }\n' \ ' }\n' \ '}\n' def and_(self, *commands): return u' -and '.join('({0})'.format(c) for c in commands) def how_to_configure(self): return { 'content': 'iex "thefuck --alias"', 'path': '$profile', 'reload': '& $profile', }
Update PowerShell alias to handle no history
Update PowerShell alias to handle no history If history is cleared (or the shell is new and there is no history), invoking thefuck results in an error because the alias attempts to execute the usage string. The fix is to check if Get-History returns anything before invoking thefuck.
Python
mit
Clpsplug/thefuck,scorphus/thefuck,nvbn/thefuck,nvbn/thefuck,scorphus/thefuck,mlk/thefuck,Clpsplug/thefuck,mlk/thefuck,SimenB/thefuck,SimenB/thefuck
--- +++ @@ -3,11 +3,14 @@ class Powershell(Generic): def app_alias(self, fuck): - return 'function ' + fuck + ' { \n' \ - ' $fuck = $(thefuck (Get-History -Count 1).CommandLine);\n' \ - ' if (-not [string]::IsNullOrWhiteSpace($fuck)) {\n' \ - ' if ($fuck.StartsWith("echo")) { $fuck = $fuck.Substring(5); }\n' \ - ' else { iex "$fuck"; }\n' \ + return 'function ' + fuck + ' {\n' \ + ' $history = (Get-History -Count 1).CommandLine;\n' \ + ' if (-not [string]::IsNullOrWhiteSpace($history)) {\n' \ + ' $fuck = $(thefuck $history);\n' \ + ' if (-not [string]::IsNullOrWhiteSpace($fuck)) {\n' \ + ' if ($fuck.StartsWith("echo")) { $fuck = $fuck.Substring(5); }\n' \ + ' else { iex "$fuck"; }\n' \ + ' }\n' \ ' }\n' \ '}\n'
7654d9dcebb0ad1e862e376b5b694234173289ed
twitter_helper/util.py
twitter_helper/util.py
import random def random_line(afile, max_chars = 123, min_chars = 5): line = next(afile) for num, aline in enumerate(afile): aline = aline.strip() if (len(aline) < min_chars or aline[0].islower() or len(aline) > max_chars) or random.randrange(num + 2): continue line = aline return line def prepare_quote(text_file, signature=" -- Hamlet", max_chars = 123, min_chars = 5,): line = random_line(text_file, max_chars, min_chars) number = random.randrange(1,1000,2) line = "{0}] " + line + signature line = line.format(number) return line
import random def random_line(afile, max_chars = 123, min_chars = 5): line = next(afile) for num, aline in enumerate(afile): aline = aline.strip() if (len(aline) < min_chars or aline[0].islower() or len(aline) > max_chars) or random.randrange(num + 2): continue line = aline #Be polite, put things back in the place you found them afile.seek(0) return line def prepare_quote(text_file, signature=" -- Hamlet", max_chars = 123, min_chars = 5,): line = random_line(text_file, max_chars, min_chars) number = random.randrange(1,1000,2) line = "{0}] " + line + signature line = line.format(number) return line
Reset pointer to the beginning of file once read it
Reset pointer to the beginning of file once read it Be polite, put things back in the place you found them
Python
mit
kuzeko/Twitter-Importer,kuzeko/Twitter-Importer
--- +++ @@ -7,6 +7,8 @@ if (len(aline) < min_chars or aline[0].islower() or len(aline) > max_chars) or random.randrange(num + 2): continue line = aline + #Be polite, put things back in the place you found them + afile.seek(0) return line def prepare_quote(text_file, signature=" -- Hamlet", max_chars = 123, min_chars = 5,):
8307188a5cbaf0dab824b58a6436affdea1b039b
mesonwrap/inventory.py
mesonwrap/inventory.py
_ORGANIZATION = 'mesonbuild' _RESTRICTED_PROJECTS = [ 'meson', 'meson-ci', 'mesonwrap', 'wrapweb', ] _RESTRICTED_ORG_PROJECTS = [ _ORGANIZATION + '/' + proj for proj in _RESTRICTED_PROJECTS ] def is_wrap_project_name(project: str) -> bool: return project not in _RESTRICTED_PROJECTS def is_wrap_full_project_name(full_project: str) -> bool: return full_project not in _RESTRICTED_ORG_PROJECTS
_ORGANIZATION = 'mesonbuild' _RESTRICTED_PROJECTS = [ 'meson', 'meson-ci', 'mesonwrap', 'wrapdevtools', 'wrapweb', ] _RESTRICTED_ORG_PROJECTS = [ _ORGANIZATION + '/' + proj for proj in _RESTRICTED_PROJECTS ] def is_wrap_project_name(project: str) -> bool: return project not in _RESTRICTED_PROJECTS def is_wrap_full_project_name(full_project: str) -> bool: return full_project not in _RESTRICTED_ORG_PROJECTS
Add wrapdevtools to restricted projects
Add wrapdevtools to restricted projects
Python
apache-2.0
mesonbuild/wrapweb,mesonbuild/wrapweb,mesonbuild/wrapweb
--- +++ @@ -3,6 +3,7 @@ 'meson', 'meson-ci', 'mesonwrap', + 'wrapdevtools', 'wrapweb', ] _RESTRICTED_ORG_PROJECTS = [
82f2fb3c3956e4ad4c65b03b3918ea409593d4ef
gcloud/__init__.py
gcloud/__init__.py
"""GCloud API access in idiomatic Python.""" __version__ = '0.02.2'
"""GCloud API access in idiomatic Python.""" from pkg_resources import get_distribution __version__ = get_distribution('gcloud').version
Read module version from setup.py
Read module version from setup.py
Python
apache-2.0
googleapis/google-cloud-python,blowmage/gcloud-python,thesandlord/gcloud-python,calpeyser/google-cloud-python,CyrusBiotechnology/gcloud-python,waprin/gcloud-python,VitalLabs/gcloud-python,waprin/google-cloud-python,jonparrott/google-cloud-python,Fkawala/gcloud-python,tswast/google-cloud-python,waprin/gcloud-python,dhermes/gcloud-python,quom/google-cloud-python,tseaver/google-cloud-python,tartavull/google-cloud-python,optimizely/gcloud-python,lucemia/gcloud-python,CyrusBiotechnology/gcloud-python,tswast/google-cloud-python,VitalLabs/gcloud-python,GrimDerp/gcloud-python,tseaver/gcloud-python,blowmage/gcloud-python,vj-ug/gcloud-python,jbuberel/gcloud-python,tswast/google-cloud-python,jonparrott/google-cloud-python,tseaver/google-cloud-python,tartavull/google-cloud-python,dhermes/google-cloud-python,tseaver/google-cloud-python,dhermes/google-cloud-python,quom/google-cloud-python,tseaver/gcloud-python,optimizely/gcloud-python,jonparrott/gcloud-python,dhermes/gcloud-python,googleapis/google-cloud-python,EugenePig/gcloud-python,EugenePig/gcloud-python,GrimDerp/gcloud-python,lucemia/gcloud-python,daspecster/google-cloud-python,jonparrott/gcloud-python,daspecster/google-cloud-python,jgeewax/gcloud-python,GoogleCloudPlatform/gcloud-python,jbuberel/gcloud-python,waprin/google-cloud-python,elibixby/gcloud-python,thesandlord/gcloud-python,GoogleCloudPlatform/gcloud-python,calpeyser/google-cloud-python,Fkawala/gcloud-python,elibixby/gcloud-python,dhermes/google-cloud-python,optimizely/gcloud-python,vj-ug/gcloud-python,jgeewax/gcloud-python
--- +++ @@ -1,4 +1,5 @@ """GCloud API access in idiomatic Python.""" +from pkg_resources import get_distribution -__version__ = '0.02.2' +__version__ = get_distribution('gcloud').version
2e88154eb9ea86bcf686e3cf4c92d5b696ec6efc
neo4j/__init__.py
neo4j/__init__.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- # Copyright (c) 2002-2016 "Neo Technology," # Network Engine for Objects in Lund AB [http://neotechnology.com] # # This file is part of Neo4j. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .meta import version as __version__
#!/usr/bin/env python # -*- encoding: utf-8 -*- # Copyright (c) 2002-2016 "Neo Technology," # Network Engine for Objects in Lund AB [http://neotechnology.com] # # This file is part of Neo4j. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from .meta import version as __version__ # Export current (v1) API. This should be updated to export the latest # version of the API when a new one is added. This gives the option to # `import neo4j.vX` for a specific version or `import neo4j` for the # latest. from .v1.constants import * from .v1.exceptions import * from .v1.session import * from .v1.types import *
Add option to import neo4j for latest version
Add option to import neo4j for latest version
Python
apache-2.0
neo4j/neo4j-python-driver,neo4j/neo4j-python-driver
--- +++ @@ -20,3 +20,12 @@ from .meta import version as __version__ + +# Export current (v1) API. This should be updated to export the latest +# version of the API when a new one is added. This gives the option to +# `import neo4j.vX` for a specific version or `import neo4j` for the +# latest. +from .v1.constants import * +from .v1.exceptions import * +from .v1.session import * +from .v1.types import *
dda54c9826b79e213432e5da1d03d171a293d42b
utils/celery_worker.py
utils/celery_worker.py
import os import sys # Append .. to sys path sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) import multiscanner from celery import Celery app = Celery('celery_worker', broker='pyamqp://guest@localhost//') @app.task def multiscanner_celery(filelist, config=multiscanner.CONFIG): ''' TODO: Add other ars + config options... This function essentially takes in a file list and runs multiscanner on them. Results are stored in the storage configured in storage.ini. Usage: from celery_worker import multiscanner_celery multiscanner_celery.delay([list, of, files, to, scan]) ''' storage_conf = multiscanner.common.get_storage_config_path(config) storage_handler = multiscanner.storage.StorageHandler(configfile=storage_conf) resultlist = multiscanner.multiscan(filelist, configfile=config) results = multiscanner.parse_reports(resultlist, python=True) storage_handler.store(results, wait=False) storage_handler.close() return results
import os import sys # Append .. to sys path sys.path.append(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) import multiscanner from celery import Celery RABBIT_USER = 'guest' RABBIT_HOST = 'localhost' app = Celery('celery_worker', broker='pyamqp://%s@%s//' % (RABBIT_USER, RABBIT_HOST)) @app.task def multiscanner_celery(filelist, config=multiscanner.CONFIG): ''' TODO: Figure out how to do batching. TODO: Add other ars + config options... This function essentially takes in a file list and runs multiscanner on them. Results are stored in the storage configured in storage.ini. Usage: from celery_worker import multiscanner_celery multiscanner_celery.delay([list, of, files, to, scan]) ''' storage_conf = multiscanner.common.get_storage_config_path(config) storage_handler = multiscanner.storage.StorageHandler(configfile=storage_conf) resultlist = multiscanner.multiscan(filelist, configfile=config) results = multiscanner.parse_reports(resultlist, python=True) storage_handler.store(results, wait=False) storage_handler.close() return results
Move rabbit vars to globals
Move rabbit vars to globals
Python
mpl-2.0
MITRECND/multiscanner,mitre/multiscanner,mitre/multiscanner,jmlong1027/multiscanner,mitre/multiscanner,jmlong1027/multiscanner,awest1339/multiscanner,MITRECND/multiscanner,awest1339/multiscanner,awest1339/multiscanner,jmlong1027/multiscanner,awest1339/multiscanner,jmlong1027/multiscanner
--- +++ @@ -6,11 +6,15 @@ from celery import Celery -app = Celery('celery_worker', broker='pyamqp://guest@localhost//') +RABBIT_USER = 'guest' +RABBIT_HOST = 'localhost' + +app = Celery('celery_worker', broker='pyamqp://%s@%s//' % (RABBIT_USER, RABBIT_HOST)) @app.task def multiscanner_celery(filelist, config=multiscanner.CONFIG): ''' + TODO: Figure out how to do batching. TODO: Add other ars + config options... This function essentially takes in a file list and runs multiscanner on them. Results are stored in the
205682120cfa77aca2b279e2ee87065e489b5e69
settings_example.py
settings_example.py
""" Example settings module. This should be copied as `settings.py` and the values modified there. That file is ignored by the repo, since it will contain environment specific and sensitive information (like passwords). """ # TODO: Allow separate settings for different subject matches. # Email formats and CSV names may change over the years, and this could # be detected by subject matches. import logging import os import re import yaml from imap import EmailCheckError, EmailServer from postgresql import DatabaseServer # If this is set to a valid path, all CSV files extracted from emails will be # stored in sub-folders within it. CSV_FOLDER = os.getcwd() SETTINGS_YAML_PATH = os.path.join(os.getcwd(), 'settings.yaml') LOGGING_FORMAT = ''' - file: %(pathname)s level: %(levelname)s line: %(lineno)s message: | %(message)s time: %(asctime)s '''.strip() LOGGING_LEVEL = logging.DEBUG def get_csv_file_types(): csv_file_types = None with open(SETTINGS_YAML_PATH) as r: csv_file_types = yaml.load(r) return csv_file_types def get_database_client(): con = 'my_username/my_password@database.example.com:5432/my_database' return DatabaseServer(con) def get_email_client(): return EmailServer('mail.example.com', 'my_username', 'my_password')
""" Example settings module. This should be copied as `settings.py` and the values modified there. That file is ignored by the repo, since it will contain environment specific and sensitive information (like passwords). """ import logging import os import re import yaml from imap import EmailCheckError, EmailServer from postgresql import DatabaseServer # If this is set to a valid path, all CSV files extracted from emails will be # stored in sub-folders within it. CSV_FOLDER = os.getcwd() SETTINGS_YAML_PATH = os.path.join(os.getcwd(), 'settings.yaml') LOGGING_FORMAT = ''' - file: %(pathname)s level: %(levelname)s line: %(lineno)s message: | %(message)s time: %(asctime)s '''.strip() LOGGING_LEVEL = logging.DEBUG def get_csv_file_types(): csv_file_types = None with open(SETTINGS_YAML_PATH) as r: csv_file_types = yaml.load(r) return csv_file_types def get_database_client(): con = 'my_username/my_password@database.example.com:5432/my_database' return DatabaseServer(con) def get_email_client(): return EmailServer('mail.example.com', 'my_username', 'my_password')
Remove TODO from settings example.
Remove TODO from settings example. This work has been started with the Yaml settings file.
Python
mit
AustralianAntarcticDataCentre/save_emails_to_files,AustralianAntarcticDataCentre/save_emails_to_files
--- +++ @@ -6,10 +6,6 @@ That file is ignored by the repo, since it will contain environment specific and sensitive information (like passwords). """ - -# TODO: Allow separate settings for different subject matches. -# Email formats and CSV names may change over the years, and this could -# be detected by subject matches. import logging import os
df3ab8bcae326ceb157106d076eaa90f13717107
astroquery/astrometry_net/tests/setup_package.py
astroquery/astrometry_net/tests/setup_package.py
# Licensed under a 3-clause BSD style license - see LICENSE.rst import os # setup paths to the test data # can specify a single file or a list of files def get_package_data(): paths = [os.path.join('data', '*.fit')] # finally construct and return a dict for the sub module return {'astroquery.astrometry_net.tests': paths}
# Licensed under a 3-clause BSD style license - see LICENSE.rst import os # setup paths to the test data # can specify a single file or a list of files def get_package_data(): paths = [os.path.join('data', '*.fit')] + [os.path.join('data', '*.fit.gz')] # finally construct and return a dict for the sub module return {'astroquery.astrometry_net.tests': paths}
Include gzipped fits files in test data
Include gzipped fits files in test data
Python
bsd-3-clause
imbasimba/astroquery,ceb8/astroquery,ceb8/astroquery,imbasimba/astroquery
--- +++ @@ -5,6 +5,6 @@ def get_package_data(): - paths = [os.path.join('data', '*.fit')] + paths = [os.path.join('data', '*.fit')] + [os.path.join('data', '*.fit.gz')] # finally construct and return a dict for the sub module return {'astroquery.astrometry_net.tests': paths}
cee291799e9aa19e23593b3618a45f7cee16d0ed
modules/cah/CAHGame.py
modules/cah/CAHGame.py
#Cards Against Humanity game engine class CAHGame: def __init__(self): self.status = "Loaded CAHGame." #flag to keep track of whether or not game is running self.running = False #list of active players in a game self.players = [] #dummy with a small deck for testing. #replace with actual card loading from DB later self.deck ={ 'questions' : [ "_? There's an app for that.", "I got 99 problems but _ ain't one.", ], 'answers' : [ "Flying sex snakes.", "Michelle Obama's arms.", "German dungeon porn.", "White people.", "Getting so angry that you pop a boner.", "Freedom of Speech", ] } #add a new player to the game def add_player(self, name): self.players.append(Player(name)) #start the game def start(self): pass def draw(self, color): '''Draws a random card of <color> from the databse and returns a Card object.''' pass #Utility class to manage Players class Player: def __init__(self, name): self.name = name #Player name (IRC nick) self.score = 0 self.hand = {} self.isCzar = False #Utiliy class for a Card class Card: def __init__(self, color, body): self.color = color self.body = body
#Cards Against Humanity game engine from cards import Deck, NoMoreCards class CAHGame: def __init__(self): self.status = "Loaded CAHGame." #flag to keep track of whether or not game is running self.running = False #list of active players in a game self.players = [] #dummy with a small deck for testing. #replace with actual card loading from DB later self.deck = Deck() #add a new player to the game def add_player(self, name): self.players.append(Player(name)) #start the game def start(self): pass #Utility class to manage Players class Player: def __init__(self, name): self.name = name #Player name (IRC nick) self.score = 0 self.hand = {} self.isCzar = False
Use the new Deck class
Use the new Deck class
Python
mit
tcoppi/scrappy,tcoppi/scrappy,johnmiked15/scrappy,johnmiked15/scrappy
--- +++ @@ -1,4 +1,6 @@ #Cards Against Humanity game engine + +from cards import Deck, NoMoreCards class CAHGame: def __init__(self): @@ -13,33 +15,14 @@ #dummy with a small deck for testing. #replace with actual card loading from DB later - self.deck ={ - 'questions' : [ - "_? There's an app for that.", - "I got 99 problems but _ ain't one.", - ], - 'answers' : [ - "Flying sex snakes.", - "Michelle Obama's arms.", - "German dungeon porn.", - "White people.", - "Getting so angry that you pop a boner.", - "Freedom of Speech", - ] - } - - + self.deck = Deck() + #add a new player to the game def add_player(self, name): self.players.append(Player(name)) #start the game def start(self): - pass - - - def draw(self, color): - '''Draws a random card of <color> from the databse and returns a Card object.''' pass @@ -50,9 +33,3 @@ self.score = 0 self.hand = {} self.isCzar = False - -#Utiliy class for a Card -class Card: - def __init__(self, color, body): - self.color = color - self.body = body
962fd486afe25031d5fb6332f623e970b694b321
tsstats/tests/test_config.py
tsstats/tests/test_config.py
import pytest from tsstats.config import load @pytest.fixture def config(): return load() def test_config(config): assert not config.getboolean('General', 'debug') assert config.getboolean('General', 'onlinedc') config.set('General', 'idmap', 'tsstats/tests/res/id_map.json') assert config.get('General', 'idmap') ==\ 'tsstats/tests/res/id_map.json' config.set('General', 'log', 'tsstats/tests/res/test.log') assert config.get('General', 'log') == 'tsstats/tests/res/test.log' config.set('General', 'output', 'output.html') assert config.get('General', 'output') == 'output.html'
import pytest from tsstats.config import load @pytest.fixture def config(): return load() def test_config(config): assert not config.getboolean('General', 'debug') assert config.getboolean('General', 'onlinedc') config.set('General', 'idmap', 'tsstats/tests/res/id_map.json') assert config.get('General', 'idmap') ==\ 'tsstats/tests/res/id_map.json' config.set('General', 'log', 'tsstats/tests/res/test.log') assert config.get('General', 'log') == 'tsstats/tests/res/test.log' config.set('General', 'output', 'output.html') assert config.get('General', 'output') == 'output.html' def test_read(): config = load(path='tsstats/tests/res/config.ini') # test defaults assert not config.getboolean('General', 'debug') # test written values assert config.get('General', 'log') == 'tsstats/tests/res/test.log' assert config.get('General', 'output') == 'tsstats/tests/res/output.html'
Test reading config from disk again
Test reading config from disk again
Python
mit
Thor77/TeamspeakStats,Thor77/TeamspeakStats
--- +++ @@ -18,3 +18,12 @@ assert config.get('General', 'log') == 'tsstats/tests/res/test.log' config.set('General', 'output', 'output.html') assert config.get('General', 'output') == 'output.html' + + +def test_read(): + config = load(path='tsstats/tests/res/config.ini') + # test defaults + assert not config.getboolean('General', 'debug') + # test written values + assert config.get('General', 'log') == 'tsstats/tests/res/test.log' + assert config.get('General', 'output') == 'tsstats/tests/res/output.html'
a2f4b30cab3dafe119e42181772f4d77b575ec0e
05/test_find_password.py
05/test_find_password.py
import unittest from find_password import find_password class TestFindPassword(unittest.TestCase): def test_find_password(self): assert find_password('abc', length=8) == '18f47a30'
import unittest from find_password import find_password class TestFindPassword(unittest.TestCase): def test_find_password(self): assert find_password('abc', length=8) == '18f47a30' assert find_password('abc', length=8, complex=True) == '05ace8e3'
Add test for part 2 of day 5.
Add test for part 2 of day 5.
Python
mit
machinelearningdeveloper/aoc_2016
--- +++ @@ -6,3 +6,4 @@ class TestFindPassword(unittest.TestCase): def test_find_password(self): assert find_password('abc', length=8) == '18f47a30' + assert find_password('abc', length=8, complex=True) == '05ace8e3'
6943bb0c665cd40e7516b7277fe55af95b814ccb
playa/conf.py
playa/conf.py
""" Represents the default values for all Sentry settings. """ import logging import os import os.path class PlayaConfig(object): ROOT = os.path.normpath(os.path.dirname(__file__)) DEBUG = True SQLITE3_DATABASE = os.path.join(ROOT, 'playa.db') AUDIO_PATHS = ['/Volumes/Storage/Music/iTunes/iTunes Media/Music/Blink-182/'] WEB_HOST = '0.0.0.0' WEB_PORT = 9000 WEB_LOG_FILE = os.path.join(ROOT, 'playa.log') WEB_PID_FILE = os.path.join(ROOT, 'playa.pid')
""" Represents the default values for all Sentry settings. """ import logging import os import os.path class PlayaConfig(object): ROOT = os.path.normpath(os.path.dirname(__file__)) DEBUG = True SQLITE3_DATABASE = os.path.join(ROOT, 'playa.db') AUDIO_PATHS = [] WEB_HOST = '0.0.0.0' WEB_PORT = 9000 WEB_LOG_FILE = os.path.join(ROOT, 'playa.log') WEB_PID_FILE = os.path.join(ROOT, 'playa.pid')
Remove my awesome default audio path
Remove my awesome default audio path
Python
apache-2.0
disqus/playa,disqus/playa
--- +++ @@ -13,7 +13,7 @@ SQLITE3_DATABASE = os.path.join(ROOT, 'playa.db') - AUDIO_PATHS = ['/Volumes/Storage/Music/iTunes/iTunes Media/Music/Blink-182/'] + AUDIO_PATHS = [] WEB_HOST = '0.0.0.0' WEB_PORT = 9000
e64101e31fadaf54f8c1d7a6acb9b302060efefc
script/lib/config.py
script/lib/config.py
#!/usr/bin/env python import platform import sys BASE_URL = 'http://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'c01b10faf0d478e48f537210ec263fabd551578d' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] verbose_mode = False def enable_verbose_mode(): print 'Running in verbose mode' global verbose_mode verbose_mode = True def is_verbose_mode(): return verbose_mode
#!/usr/bin/env python import platform import sys BASE_URL = 'http://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' LIBCHROMIUMCONTENT_COMMIT = 'e0213676879061470efe50720368bce9b99aaa12' ARCH = { 'cygwin': '32bit', 'darwin': '64bit', 'linux2': platform.architecture()[0], 'win32': '32bit', }[sys.platform] DIST_ARCH = { '32bit': 'ia32', '64bit': 'x64', }[ARCH] TARGET_PLATFORM = { 'cygwin': 'win32', 'darwin': 'darwin', 'linux2': 'linux', 'win32': 'win32', }[sys.platform] verbose_mode = False def enable_verbose_mode(): print 'Running in verbose mode' global verbose_mode verbose_mode = True def is_verbose_mode(): return verbose_mode
Upgrade libchromiumcontent to use the static_library build
Upgrade libchromiumcontent to use the static_library build
Python
mit
wan-qy/electron,deed02392/electron,John-Lin/electron,aliib/electron,nicobot/electron,Andrey-Pavlov/electron,leethomas/electron,Jonekee/electron,simongregory/electron,leolujuyi/electron,michaelchiche/electron,greyhwndz/electron,sircharleswatson/electron,jacksondc/electron,jsutcodes/electron,bpasero/electron,tylergibson/electron,christian-bromann/electron,egoist/electron,medixdev/electron,howmuchcomputer/electron,egoist/electron,michaelchiche/electron,kenmozi/electron,brave/muon,smczk/electron,seanchas116/electron,ianscrivener/electron,Zagorakiss/electron,pandoraui/electron,renaesop/electron,shiftkey/electron,farmisen/electron,astoilkov/electron,nicobot/electron,thomsonreuters/electron,beni55/electron,setzer777/electron,vHanda/electron,Zagorakiss/electron,JussMee15/electron,christian-bromann/electron,SufianHassan/electron,shaundunne/electron,fomojola/electron,aichingm/electron,webmechanicx/electron,the-ress/electron,neutrous/electron,chriskdon/electron,pombredanne/electron,zhakui/electron,etiktin/electron,smczk/electron,dongjoon-hyun/electron,iftekeriba/electron,tylergibson/electron,Rokt33r/electron,vipulroxx/electron,kazupon/electron,egoist/electron,fritx/electron,digideskio/electron,bobwol/electron,bright-sparks/electron,davazp/electron,rajatsingla28/electron,beni55/electron,mattotodd/electron,abhishekgahlot/electron,ankitaggarwal011/electron,JussMee15/electron,kenmozi/electron,meowlab/electron,MaxGraey/electron,tonyganch/electron,miniak/electron,kcrt/electron,egoist/electron,roadev/electron,Zagorakiss/electron,iftekeriba/electron,ervinb/electron,micalan/electron,jcblw/electron,mubassirhayat/electron,carsonmcdonald/electron,rreimann/electron,yan-foto/electron,timruffles/electron,JesselJohn/electron,tonyganch/electron,thompsonemerson/electron,destan/electron,gerhardberger/electron,icattlecoder/electron,brave/electron,sircharleswatson/electron,noikiy/electron,lrlna/electron,takashi/electron,vipulroxx/electron,sshiting/electron,systembugtj/electron,beni55/electron,bpasero/electron,jacksondc/electron,kenmozi/electron,Jacobichou/electron,carsonmcdonald/electron,voidbridge/electron,smczk/electron,trigrass2/electron,tinydew4/electron,noikiy/electron,soulteary/electron,aliib/electron,biblerule/UMCTelnetHub,jhen0409/electron,the-ress/electron,jjz/electron,bpasero/electron,mirrh/electron,natgolov/electron,hokein/atom-shell,Jonekee/electron,gamedevsam/electron,saronwei/electron,rsvip/electron,pandoraui/electron,jannishuebl/electron,gabriel/electron,electron/electron,simonfork/electron,MaxWhere/electron,eric-seekas/electron,gerhardberger/electron,etiktin/electron,thomsonreuters/electron,bright-sparks/electron,biblerule/UMCTelnetHub,preco21/electron,BionicClick/electron,Jacobichou/electron,sshiting/electron,jtburke/electron,christian-bromann/electron,brave/muon,icattlecoder/electron,astoilkov/electron,jiaz/electron,miniak/electron,shockone/electron,shennushi/electron,Ivshti/electron,brenca/electron,Gerhut/electron,anko/electron,tinydew4/electron,voidbridge/electron,setzer777/electron,trankmichael/electron,simongregory/electron,pirafrank/electron,mrwizard82d1/electron,joaomoreno/atom-shell,jlhbaseball15/electron,jcblw/electron,simongregory/electron,Jonekee/electron,kazupon/electron,miniak/electron,adamjgray/electron,farmisen/electron,medixdev/electron,minggo/electron,bbondy/electron,twolfson/electron,roadev/electron,nekuz0r/electron,oiledCode/electron,davazp/electron,brenca/electron,sircharleswatson/electron,zhakui/electron,ianscrivener/electron,vipulroxx/electron,jhen0409/electron,coderhaoxin/electron,setzer777/electron,vaginessa/electron,eriser/electron,Gerhut/electron,cqqccqc/electron,jiaz/electron,Andrey-Pavlov/electron,electron/electron,jsutcodes/electron,IonicaBizauKitchen/electron,abhishekgahlot/electron,astoilkov/electron,arusakov/electron,chriskdon/electron,evgenyzinoviev/electron,bitemyapp/electron,xfstudio/electron,rajatsingla28/electron,hokein/atom-shell,yan-foto/electron,cqqccqc/electron,deed02392/electron,sshiting/electron,egoist/electron,jiaz/electron,arusakov/electron,davazp/electron,saronwei/electron,kokdemo/electron,mattotodd/electron,Faiz7412/electron,leolujuyi/electron,takashi/electron,Evercoder/electron,jaanus/electron,Evercoder/electron,soulteary/electron,leethomas/electron,miniak/electron,JussMee15/electron,sircharleswatson/electron,cqqccqc/electron,fffej/electron,jannishuebl/electron,minggo/electron,kcrt/electron,aichingm/electron,saronwei/electron,SufianHassan/electron,brave/electron,rsvip/electron,evgenyzinoviev/electron,kostia/electron,thomsonreuters/electron,aliib/electron,hokein/atom-shell,shennushi/electron,d-salas/electron,trigrass2/electron,ankitaggarwal011/electron,cos2004/electron,nicobot/electron,bpasero/electron,adcentury/electron,yalexx/electron,tincan24/electron,pandoraui/electron,jcblw/electron,fritx/electron,mrwizard82d1/electron,hokein/atom-shell,icattlecoder/electron,webmechanicx/electron,astoilkov/electron,bbondy/electron,mattdesl/electron,jtburke/electron,Floato/electron,bruce/electron,Andrey-Pavlov/electron,eric-seekas/electron,d-salas/electron,IonicaBizauKitchen/electron,LadyNaggaga/electron,stevemao/electron,bright-sparks/electron,benweissmann/electron,Rokt33r/electron,bitemyapp/electron,tinydew4/electron,leftstick/electron,tylergibson/electron,thingsinjars/electron,anko/electron,preco21/electron,kcrt/electron,pandoraui/electron,digideskio/electron,takashi/electron,Rokt33r/electron,mirrh/electron,shockone/electron,meowlab/electron,shaundunne/electron,rsvip/electron,arusakov/electron,LadyNaggaga/electron,lrlna/electron,jsutcodes/electron,MaxGraey/electron,RIAEvangelist/electron,nicobot/electron,gerhardberger/electron,fffej/electron,shennushi/electron,trankmichael/electron,gabrielPeart/electron,wan-qy/electron,robinvandernoord/electron,destan/electron,fomojola/electron,thomsonreuters/electron,medixdev/electron,edulan/electron,ervinb/electron,tincan24/electron,arturts/electron,xiruibing/electron,Zagorakiss/electron,meowlab/electron,ankitaggarwal011/electron,coderhaoxin/electron,jonatasfreitasv/electron,dkfiresky/electron,brave/muon,dongjoon-hyun/electron,voidbridge/electron,miniak/electron,leftstick/electron,DivyaKMenon/electron,wan-qy/electron,pirafrank/electron,mattotodd/electron,rreimann/electron,dongjoon-hyun/electron,seanchas116/electron,gabriel/electron,trigrass2/electron,DivyaKMenon/electron,sircharleswatson/electron,farmisen/electron,xiruibing/electron,bobwol/electron,d-salas/electron,matiasinsaurralde/electron,Gerhut/electron,minggo/electron,adcentury/electron,yan-foto/electron,micalan/electron,seanchas116/electron,deed02392/electron,wolfflow/electron,thingsinjars/electron,shiftkey/electron,Gerhut/electron,leftstick/electron,iftekeriba/electron,rreimann/electron,MaxWhere/electron,medixdev/electron,Floato/electron,shockone/electron,twolfson/electron,LadyNaggaga/electron,sshiting/electron,chriskdon/electron,baiwyc119/electron,smczk/electron,ervinb/electron,mhkeller/electron,fritx/electron,coderhaoxin/electron,jhen0409/electron,thompsonemerson/electron,pombredanne/electron,noikiy/electron,jhen0409/electron,faizalpribadi/electron,LadyNaggaga/electron,bobwol/electron,twolfson/electron,mirrh/electron,eric-seekas/electron,tonyganch/electron,brave/electron,sky7sea/electron,thingsinjars/electron,pombredanne/electron,edulan/electron,RIAEvangelist/electron,Ivshti/electron,mhkeller/electron,kazupon/electron,joaomoreno/atom-shell,RobertJGabriel/electron,natgolov/electron,benweissmann/electron,Rokt33r/electron,pombredanne/electron,setzer777/electron,gabriel/electron,cos2004/electron,stevemao/electron,fabien-d/electron,pirafrank/electron,fomojola/electron,shaundunne/electron,baiwyc119/electron,mjaniszew/electron,darwin/electron,ervinb/electron,BionicClick/electron,thomsonreuters/electron,Floato/electron,evgenyzinoviev/electron,coderhaoxin/electron,Rokt33r/electron,lrlna/electron,benweissmann/electron,SufianHassan/electron,bobwol/electron,bbondy/electron,rajatsingla28/electron,gamedevsam/electron,nicholasess/electron,roadev/electron,synaptek/electron,electron/electron,christian-bromann/electron,bitemyapp/electron,rreimann/electron,greyhwndz/electron,lrlna/electron,zhakui/electron,neutrous/electron,systembugtj/electron,tinydew4/electron,leethomas/electron,SufianHassan/electron,jsutcodes/electron,gerhardberger/electron,rhencke/electron,setzer777/electron,gbn972/electron,trankmichael/electron,rajatsingla28/electron,pombredanne/electron,Evercoder/electron,edulan/electron,farmisen/electron,JesselJohn/electron,mjaniszew/electron,jacksondc/electron,vipulroxx/electron,shiftkey/electron,abhishekgahlot/electron,meowlab/electron,JussMee15/electron,ankitaggarwal011/electron,minggo/electron,Floato/electron,farmisen/electron,aaron-goshine/electron,jaanus/electron,dongjoon-hyun/electron,jlhbaseball15/electron,LadyNaggaga/electron,jannishuebl/electron,fomojola/electron,RobertJGabriel/electron,oiledCode/electron,joaomoreno/atom-shell,tomashanacek/electron,bwiggs/electron,soulteary/electron,bpasero/electron,fabien-d/electron,trigrass2/electron,joaomoreno/atom-shell,preco21/electron,timruffles/electron,jcblw/electron,fomojola/electron,yan-foto/electron,deed02392/electron,DivyaKMenon/electron,seanchas116/electron,jaanus/electron,aecca/electron,subblue/electron,miniak/electron,synaptek/electron,bwiggs/electron,vaginessa/electron,John-Lin/electron,GoooIce/electron,meowlab/electron,wan-qy/electron,bruce/electron,jannishuebl/electron,wolfflow/electron,Neron-X5/electron,davazp/electron,tincan24/electron,renaesop/electron,jonatasfreitasv/electron,ankitaggarwal011/electron,d-salas/electron,abhishekgahlot/electron,subblue/electron,jacksondc/electron,tomashanacek/electron,IonicaBizauKitchen/electron,Jacobichou/electron,aecca/electron,yalexx/electron,leolujuyi/electron,Floato/electron,Jonekee/electron,noikiy/electron,timruffles/electron,IonicaBizauKitchen/electron,jiaz/electron,brave/muon,meowlab/electron,gerhardberger/electron,gerhardberger/electron,nekuz0r/electron,felixrieseberg/electron,dkfiresky/electron,timruffles/electron,dahal/electron,minggo/electron,felixrieseberg/electron,stevekinney/electron,eriser/electron,the-ress/electron,joneit/electron,xfstudio/electron,mattdesl/electron,oiledCode/electron,smczk/electron,trigrass2/electron,dahal/electron,kenmozi/electron,mattdesl/electron,bright-sparks/electron,chriskdon/electron,voidbridge/electron,adcentury/electron,aaron-goshine/electron,cos2004/electron,vHanda/electron,setzer777/electron,voidbridge/electron,subblue/electron,sky7sea/electron,sky7sea/electron,preco21/electron,dkfiresky/electron,faizalpribadi/electron,benweissmann/electron,rsvip/electron,howmuchcomputer/electron,micalan/electron,yan-foto/electron,neutrous/electron,mhkeller/electron,Evercoder/electron,joaomoreno/atom-shell,eric-seekas/electron,tincan24/electron,jlord/electron,John-Lin/electron,adamjgray/electron,dongjoon-hyun/electron,the-ress/electron,RobertJGabriel/electron,chrisswk/electron,cqqccqc/electron,posix4e/electron,John-Lin/electron,tomashanacek/electron,kenmozi/electron,rhencke/electron,wan-qy/electron,aecca/electron,RobertJGabriel/electron,kostia/electron,tinydew4/electron,fffej/electron,rhencke/electron,nicholasess/electron,etiktin/electron,carsonmcdonald/electron,cos2004/electron,d-salas/electron,jcblw/electron,destan/electron,jacksondc/electron,thompsonemerson/electron,aaron-goshine/electron,anko/electron,greyhwndz/electron,GoooIce/electron,robinvandernoord/electron,rsvip/electron,faizalpribadi/electron,icattlecoder/electron,twolfson/electron,systembugtj/electron,vipulroxx/electron,leethomas/electron,rhencke/electron,electron/electron,fritx/electron,bruce/electron,Rokt33r/electron,shockone/electron,shockone/electron,bpasero/electron,bitemyapp/electron,gbn972/electron,BionicClick/electron,aaron-goshine/electron,Jacobichou/electron,matiasinsaurralde/electron,chriskdon/electron,nicholasess/electron,leftstick/electron,howmuchcomputer/electron,edulan/electron,bruce/electron,nicholasess/electron,MaxGraey/electron,Faiz7412/electron,vHanda/electron,simongregory/electron,adamjgray/electron,joneit/electron,digideskio/electron,JussMee15/electron,BionicClick/electron,renaesop/electron,howmuchcomputer/electron,rajatsingla28/electron,cos2004/electron,nicobot/electron,brave/electron,jjz/electron,robinvandernoord/electron,destan/electron,fabien-d/electron,brenca/electron,jaanus/electron,natgolov/electron,brenca/electron,IonicaBizauKitchen/electron,tincan24/electron,shennushi/electron,astoilkov/electron,adcentury/electron,gabrielPeart/electron,Gerhut/electron,leftstick/electron,evgenyzinoviev/electron,simonfork/electron,SufianHassan/electron,christian-bromann/electron,mattdesl/electron,tonyganch/electron,stevekinney/electron,michaelchiche/electron,jannishuebl/electron,shaundunne/electron,webmechanicx/electron,systembugtj/electron,gbn972/electron,xiruibing/electron,pirafrank/electron,renaesop/electron,d-salas/electron,anko/electron,minggo/electron,adcentury/electron,preco21/electron,gbn972/electron,adamjgray/electron,tomashanacek/electron,fireball-x/atom-shell,jcblw/electron,fireball-x/atom-shell,vipulroxx/electron,soulteary/electron,subblue/electron,jacksondc/electron,aliib/electron,thingsinjars/electron,mrwizard82d1/electron,rreimann/electron,kokdemo/electron,kostia/electron,lzpfmh/electron,MaxWhere/electron,adcentury/electron,SufianHassan/electron,aaron-goshine/electron,micalan/electron,mattotodd/electron,noikiy/electron,arusakov/electron,gabrielPeart/electron,leolujuyi/electron,dahal/electron,digideskio/electron,destan/electron,ianscrivener/electron,leolujuyi/electron,mirrh/electron,synaptek/electron,eric-seekas/electron,IonicaBizauKitchen/electron,oiledCode/electron,adamjgray/electron,aliib/electron,jiaz/electron,joneit/electron,deed02392/electron,jtburke/electron,yalexx/electron,fritx/electron,brenca/electron,aaron-goshine/electron,matiasinsaurralde/electron,howmuchcomputer/electron,bwiggs/electron,gamedevsam/electron,medixdev/electron,xfstudio/electron,Neron-X5/electron,bitemyapp/electron,GoooIce/electron,jlhbaseball15/electron,wolfflow/electron,gamedevsam/electron,Andrey-Pavlov/electron,michaelchiche/electron,neutrous/electron,Jacobichou/electron,wolfflow/electron,tylergibson/electron,bbondy/electron,bwiggs/electron,faizalpribadi/electron,saronwei/electron,roadev/electron,mattotodd/electron,BionicClick/electron,twolfson/electron,dkfiresky/electron,systembugtj/electron,Faiz7412/electron,jsutcodes/electron,aichingm/electron,oiledCode/electron,kazupon/electron,shaundunne/electron,zhakui/electron,RobertJGabriel/electron,felixrieseberg/electron,leolujuyi/electron,trigrass2/electron,thompsonemerson/electron,shennushi/electron,fabien-d/electron,gabrielPeart/electron,seanchas116/electron,darwin/electron,gabrielPeart/electron,GoooIce/electron,Faiz7412/electron,lzpfmh/electron,beni55/electron,jhen0409/electron,jlord/electron,kcrt/electron,benweissmann/electron,jonatasfreitasv/electron,pandoraui/electron,shennushi/electron,sky7sea/electron,aecca/electron,astoilkov/electron,icattlecoder/electron,gabriel/electron,brave/muon,vHanda/electron,pirafrank/electron,abhishekgahlot/electron,sshiting/electron,fireball-x/atom-shell,chrisswk/electron,jaanus/electron,felixrieseberg/electron,xfstudio/electron,chriskdon/electron,posix4e/electron,dkfiresky/electron,preco21/electron,RIAEvangelist/electron,stevemao/electron,Neron-X5/electron,Gerhut/electron,kcrt/electron,GoooIce/electron,vHanda/electron,bwiggs/electron,eric-seekas/electron,bitemyapp/electron,felixrieseberg/electron,howmuchcomputer/electron,xiruibing/electron,kokdemo/electron,mjaniszew/electron,aichingm/electron,iftekeriba/electron,evgenyzinoviev/electron,faizalpribadi/electron,mrwizard82d1/electron,stevemao/electron,takashi/electron,arturts/electron,leethomas/electron,tylergibson/electron,jlord/electron,fffej/electron,gbn972/electron,kenmozi/electron,brenca/electron,bright-sparks/electron,stevemao/electron,eriser/electron,sky7sea/electron,felixrieseberg/electron,pirafrank/electron,biblerule/UMCTelnetHub,renaesop/electron,gabriel/electron,lzpfmh/electron,simonfork/electron,bbondy/electron,lrlna/electron,michaelchiche/electron,jlhbaseball15/electron,ervinb/electron,baiwyc119/electron,kostia/electron,jiaz/electron,the-ress/electron,ianscrivener/electron,DivyaKMenon/electron,MaxGraey/electron,micalan/electron,kazupon/electron,the-ress/electron,oiledCode/electron,Zagorakiss/electron,baiwyc119/electron,jonatasfreitasv/electron,jaanus/electron,nicholasess/electron,chrisswk/electron,JussMee15/electron,John-Lin/electron,Andrey-Pavlov/electron,gamedevsam/electron,vaginessa/electron,robinvandernoord/electron,thomsonreuters/electron,leethomas/electron,medixdev/electron,chrisswk/electron,rajatsingla28/electron,davazp/electron,dahal/electron,rreimann/electron,carsonmcdonald/electron,mubassirhayat/electron,kcrt/electron,yalexx/electron,arturts/electron,chrisswk/electron,posix4e/electron,trankmichael/electron,edulan/electron,joneit/electron,nekuz0r/electron,xiruibing/electron,tomashanacek/electron,nekuz0r/electron,stevekinney/electron,gerhardberger/electron,Evercoder/electron,nicobot/electron,joneit/electron,jlord/electron,yalexx/electron,roadev/electron,stevekinney/electron,mrwizard82d1/electron,JesselJohn/electron,digideskio/electron,Evercoder/electron,simonfork/electron,vHanda/electron,brave/electron,roadev/electron,soulteary/electron,Faiz7412/electron,JesselJohn/electron,GoooIce/electron,icattlecoder/electron,gamedevsam/electron,arturts/electron,yan-foto/electron,arturts/electron,Ivshti/electron,ankitaggarwal011/electron,mubassirhayat/electron,xfstudio/electron,darwin/electron,carsonmcdonald/electron,mattotodd/electron,takashi/electron,hokein/atom-shell,jlhbaseball15/electron,lrlna/electron,MaxWhere/electron,tonyganch/electron,mjaniszew/electron,aliib/electron,darwin/electron,ervinb/electron,kokdemo/electron,mhkeller/electron,yalexx/electron,MaxWhere/electron,christian-bromann/electron,JesselJohn/electron,synaptek/electron,wolfflow/electron,beni55/electron,rhencke/electron,coderhaoxin/electron,fritx/electron,ianscrivener/electron,rhencke/electron,sshiting/electron,jannishuebl/electron,eriser/electron,vaginessa/electron,etiktin/electron,cos2004/electron,mirrh/electron,simongregory/electron,twolfson/electron,electron/electron,destan/electron,arturts/electron,synaptek/electron,sircharleswatson/electron,digideskio/electron,Neron-X5/electron,joneit/electron,thingsinjars/electron,Floato/electron,beni55/electron,jlhbaseball15/electron,mjaniszew/electron,Andrey-Pavlov/electron,tonyganch/electron,arusakov/electron,dkfiresky/electron,davazp/electron,dahal/electron,mubassirhayat/electron,mubassirhayat/electron,leftstick/electron,DivyaKMenon/electron,sky7sea/electron,joaomoreno/atom-shell,jtburke/electron,bobwol/electron,baiwyc119/electron,simongregory/electron,jhen0409/electron,natgolov/electron,deed02392/electron,coderhaoxin/electron,Zagorakiss/electron,edulan/electron,natgolov/electron,iftekeriba/electron,thompsonemerson/electron,jonatasfreitasv/electron,electron/electron,jsutcodes/electron,shiftkey/electron,MaxGraey/electron,shiftkey/electron,kostia/electron,Jonekee/electron,aecca/electron,farmisen/electron,nicholasess/electron,kazupon/electron,bpasero/electron,anko/electron,matiasinsaurralde/electron,bruce/electron,webmechanicx/electron,mattdesl/electron,gabriel/electron,aichingm/electron,biblerule/UMCTelnetHub,stevekinney/electron,JesselJohn/electron,vaginessa/electron,seanchas116/electron,John-Lin/electron,Ivshti/electron,posix4e/electron,benweissmann/electron,Neron-X5/electron,jjz/electron,eriser/electron,thompsonemerson/electron,wan-qy/electron,tinydew4/electron,jonatasfreitasv/electron,fabien-d/electron,neutrous/electron,synaptek/electron,greyhwndz/electron,MaxWhere/electron,saronwei/electron,noikiy/electron,tylergibson/electron,wolfflow/electron,jjz/electron,bobwol/electron,thingsinjars/electron,RIAEvangelist/electron,takashi/electron,fireball-x/atom-shell,micalan/electron,darwin/electron,bbondy/electron,subblue/electron,webmechanicx/electron,jjz/electron,michaelchiche/electron,Jacobichou/electron,zhakui/electron,matiasinsaurralde/electron,gbn972/electron,Ivshti/electron,fomojola/electron,renaesop/electron,shockone/electron,kokdemo/electron,etiktin/electron,subblue/electron,biblerule/UMCTelnetHub,bruce/electron,robinvandernoord/electron,posix4e/electron,zhakui/electron,mhkeller/electron,smczk/electron,etiktin/electron,iftekeriba/electron,faizalpribadi/electron,carsonmcdonald/electron,BionicClick/electron,arusakov/electron,timruffles/electron,cqqccqc/electron,Neron-X5/electron,lzpfmh/electron,eriser/electron,electron/electron,pombredanne/electron,jjz/electron,bright-sparks/electron,simonfork/electron,RIAEvangelist/electron,LadyNaggaga/electron,aecca/electron,RIAEvangelist/electron,anko/electron,mrwizard82d1/electron,greyhwndz/electron,evgenyzinoviev/electron,xiruibing/electron,mjaniszew/electron,abhishekgahlot/electron,jtburke/electron,mirrh/electron,mhkeller/electron,lzpfmh/electron,tomashanacek/electron,neutrous/electron,fffej/electron,simonfork/electron,RobertJGabriel/electron,biblerule/UMCTelnetHub,baiwyc119/electron,egoist/electron,jlord/electron,Jonekee/electron,jtburke/electron,stevemao/electron,webmechanicx/electron,voidbridge/electron,kokdemo/electron,cqqccqc/electron,vaginessa/electron,the-ress/electron,brave/electron,adamjgray/electron,gabrielPeart/electron,systembugtj/electron,matiasinsaurralde/electron,ianscrivener/electron,nekuz0r/electron,kostia/electron,trankmichael/electron,lzpfmh/electron,natgolov/electron,robinvandernoord/electron,bwiggs/electron,nekuz0r/electron,mattdesl/electron,shaundunne/electron,fffej/electron,DivyaKMenon/electron,posix4e/electron,saronwei/electron,xfstudio/electron,aichingm/electron,stevekinney/electron,shiftkey/electron,trankmichael/electron,dongjoon-hyun/electron,dahal/electron,pandoraui/electron,soulteary/electron,brave/muon,fireball-x/atom-shell,greyhwndz/electron,tincan24/electron
--- +++ @@ -4,7 +4,7 @@ import sys BASE_URL = 'http://gh-contractor-zcbenz.s3.amazonaws.com/libchromiumcontent' -LIBCHROMIUMCONTENT_COMMIT = 'c01b10faf0d478e48f537210ec263fabd551578d' +LIBCHROMIUMCONTENT_COMMIT = 'e0213676879061470efe50720368bce9b99aaa12' ARCH = { 'cygwin': '32bit',
d5a578e6b72fae3c92827895055ed32baf8aa806
coney/response_codes.py
coney/response_codes.py
class ResponseCodes(object): SUCCESS = 0 USER_CODE_START = 1 USER_CODE_END = 0x7fffffff RESERVED_CODE_START = 0x80000000 MALFORMED_RESPONSE = RESERVED_CODE_START REQUEST_ENCODING_FAILURE = RESERVED_CODE_START + 1 REMOTE_UNHANDLED_EXCEPTION = RESERVED_CODE_START + 2 CALL_REPLY_TIMEOUT = RESERVED_CODE_START + 3 RESERVED_CODE_END = 0xffffffff _desc = { SUCCESS: 'Success', MALFORMED_RESPONSE: 'Response message was malformed', REQUEST_ENCODING_FAILURE: 'The data in the request could not be encoded', REMOTE_UNHANDLED_EXCEPTION: 'An unhandled exception occurred while processing the remote call', CALL_REPLY_TIMEOUT: 'The request did not receive a reply within the call timeout', } @staticmethod def describe(code): try: return ResponseCodes._desc[code] except KeyError: if ResponseCodes.USER_CODE_START >= code <= ResponseCodes.USER_CODE_END: return 'RPC endpoint specific error response' else: return 'Unknown response code'
class ResponseCodes(object): SUCCESS = 0 USER_CODE_START = 1 USER_CODE_END = 0x7fffffff RESERVED_CODE_START = 0x80000000 MALFORMED_RESPONSE = RESERVED_CODE_START MALFORMED_REQUEST = RESERVED_CODE_START + 1 REQUEST_ENCODING_FAILURE = RESERVED_CODE_START + 2 REMOTE_UNHANDLED_EXCEPTION = RESERVED_CODE_START + 3 CALL_REPLY_TIMEOUT = RESERVED_CODE_START + 4 METHOD_NOT_FOUND = RESERVED_CODE_START + 5 VERSION_NOT_FOUND = RESERVED_CODE_START + 6 UNEXPECTED_DISPATCH_EXCEPTION = RESERVED_CODE_START + 7 RESERVED_CODE_END = 0xffffffff _desc = { SUCCESS: 'Success', MALFORMED_RESPONSE: 'Response message was malformed', MALFORMED_REQUEST: 'Request message was malformed', REQUEST_ENCODING_FAILURE: 'The data in the request could not be encoded', REMOTE_UNHANDLED_EXCEPTION: 'An unhandled exception occurred while processing the remote call', CALL_REPLY_TIMEOUT: 'The request did not receive a reply within the call timeout', METHOD_NOT_FOUND: 'The requested method is not supported by the server', VERSION_NOT_FOUND: 'The requested method version is not supported by the server', UNEXPECTED_DISPATCH_EXCEPTION: 'An unexpected exception occurred during message dispatch' } @staticmethod def describe(code): try: return ResponseCodes._desc[code] except KeyError: if ResponseCodes.USER_CODE_START >= code <= ResponseCodes.USER_CODE_END: return 'RPC endpoint specific error response' else: return 'Unknown response code'
Add additional codes used by server implementation
Add additional codes used by server implementation
Python
mit
cbigler/jackrabbit
--- +++ @@ -8,17 +8,26 @@ RESERVED_CODE_START = 0x80000000 MALFORMED_RESPONSE = RESERVED_CODE_START - REQUEST_ENCODING_FAILURE = RESERVED_CODE_START + 1 - REMOTE_UNHANDLED_EXCEPTION = RESERVED_CODE_START + 2 - CALL_REPLY_TIMEOUT = RESERVED_CODE_START + 3 + MALFORMED_REQUEST = RESERVED_CODE_START + 1 + REQUEST_ENCODING_FAILURE = RESERVED_CODE_START + 2 + REMOTE_UNHANDLED_EXCEPTION = RESERVED_CODE_START + 3 + CALL_REPLY_TIMEOUT = RESERVED_CODE_START + 4 + METHOD_NOT_FOUND = RESERVED_CODE_START + 5 + VERSION_NOT_FOUND = RESERVED_CODE_START + 6 + UNEXPECTED_DISPATCH_EXCEPTION = RESERVED_CODE_START + 7 + RESERVED_CODE_END = 0xffffffff _desc = { SUCCESS: 'Success', MALFORMED_RESPONSE: 'Response message was malformed', + MALFORMED_REQUEST: 'Request message was malformed', REQUEST_ENCODING_FAILURE: 'The data in the request could not be encoded', REMOTE_UNHANDLED_EXCEPTION: 'An unhandled exception occurred while processing the remote call', CALL_REPLY_TIMEOUT: 'The request did not receive a reply within the call timeout', + METHOD_NOT_FOUND: 'The requested method is not supported by the server', + VERSION_NOT_FOUND: 'The requested method version is not supported by the server', + UNEXPECTED_DISPATCH_EXCEPTION: 'An unexpected exception occurred during message dispatch' } @staticmethod
adc3fa70c32bce764a6b6a7efd7a39c349d3a685
quick_sort.py
quick_sort.py
"""Doc string to end all doc strings""" def quick_srt(un_list): _helper(un_list, 0, len(un_list)-1) def _helper(un_list, first, last): if first < last: split = _split(un_list, first, last) _helper(un_list, first, split-1) _helper(un_list, split+1, last) def _split(un_list, first, last): pivot = un_list[first] left = first + 1 right = last while True: while left <= right and un_list[left] <= pivot: left += 1 while right >= left and un_list[right] >= pivot: right += 1 if right < left: break else: temp = un_list[left] un_list[left] = un_list[right] un_list[right] = temp temp = un_list[first] un_list[first] = un_list[right] un_list[right] = temp return right if __name__ == '__main__': pass
"""Doc string to end all doc strings""" def quick_srt(un_list): _helper(un_list, 0, len(un_list)-1) def _helper(un_list, first, last): if first < last: split = _split(un_list, first, last) _helper(un_list, first, split-1) _helper(un_list, split+1, last) def _split(un_list, first, last): pivot = un_list[first] left = first + 1 right = last while True: while left <= right and un_list[left] <= pivot: left += 1 while un_list[right] >= pivot and right >= left: right -= 1 if right < left: break else: temp = un_list[left] un_list[left] = un_list[right] un_list[right] = temp temp = un_list[first] un_list[first] = un_list[right] un_list[right] = temp return right if __name__ == '__main__': from random import shuffle rands = [2 for num in range(0, 1001)] nums = range(0, 1001) BEST_CASE = shuffle(nums) WORST_CASE = nums from timeit import Timer SETUP = """from __main__ import BEST_CASE, WORST_CASE, quick_srt""" best = Timer('quick_srt({})'.format(BEST_CASE), SETUP).timeit(100) worst = Timer('quick_srt({})'.format(WORST_CASE), SETUP).timeit(100) print(""" Best case represented as a list that is Worst case represented as a list that is """) print('Best Case: {}'.format(best)) print('Worst Case: {}'.format(worst))
Update index error in _sort method
Update index error in _sort method
Python
mit
jonathanstallings/data-structures
--- +++ @@ -20,8 +20,8 @@ while True: while left <= right and un_list[left] <= pivot: left += 1 - while right >= left and un_list[right] >= pivot: - right += 1 + while un_list[right] >= pivot and right >= left: + right -= 1 if right < left: break @@ -38,4 +38,23 @@ if __name__ == '__main__': - pass + from random import shuffle + rands = [2 for num in range(0, 1001)] + nums = range(0, 1001) + BEST_CASE = shuffle(nums) + WORST_CASE = nums + + from timeit import Timer + + SETUP = """from __main__ import BEST_CASE, WORST_CASE, quick_srt""" + + best = Timer('quick_srt({})'.format(BEST_CASE), SETUP).timeit(100) + + worst = Timer('quick_srt({})'.format(WORST_CASE), SETUP).timeit(100) + + print(""" + Best case represented as a list that is + Worst case represented as a list that is + """) + print('Best Case: {}'.format(best)) + print('Worst Case: {}'.format(worst))
fc7aac4f68c4b694162ed146b8b5ab2b4401895c
test/features/test_create_pages.py
test/features/test_create_pages.py
import time import unittest from hamcrest import * from splinter import Browser from support.stub_server import HttpStub class test_create_pages(unittest.TestCase): def setUp(self): HttpStub.start() time.sleep(2) def tearDown(self): HttpStub.stop() def test_about_page(self): with Browser() as browser: browser.visit("http://0.0.0.0:8000/aboutData") assert_that(browser.is_text_present('About the transactions data'), is_(True))
import time import unittest from hamcrest import * from splinter import Browser from support.stub_server import HttpStub class test_create_pages(unittest.TestCase): def setUp(self): HttpStub.start() time.sleep(2) def tearDown(self): HttpStub.stop() def test_about_page(self): with Browser() as browser: browser.visit("http://0.0.0.0:8000/high-volume-services/by-transactions-per-year/descending.html") assert_that(browser.is_text_present('High-volume services'), is_(True))
Test for page existing on master branch
Test for page existing on master branch
Python
mit
alphagov/transactions-explorer,gds-attic/transactions-explorer,gds-attic/transactions-explorer,gds-attic/transactions-explorer,gds-attic/transactions-explorer,alphagov/transactions-explorer,alphagov/transactions-explorer,gds-attic/transactions-explorer,alphagov/transactions-explorer,alphagov/transactions-explorer
--- +++ @@ -18,6 +18,6 @@ def test_about_page(self): with Browser() as browser: - browser.visit("http://0.0.0.0:8000/aboutData") - assert_that(browser.is_text_present('About the transactions data'), + browser.visit("http://0.0.0.0:8000/high-volume-services/by-transactions-per-year/descending.html") + assert_that(browser.is_text_present('High-volume services'), is_(True))
209fef39f72a625e154f4455eaa6754d6a85e98b
zeus/utils/revisions.py
zeus/utils/revisions.py
from dataclasses import dataclass from typing import List, Tuple from zeus.exceptions import UnknownRevision from zeus.models import Repository, Revision from zeus.vcs import vcs_client @dataclass class RevisionResult: sha: str message: str author: str author_date: str committer: str committer_date: str parents: List[str] authors: List[Tuple[str, str]] def identify_revision( repository: Repository, ref: str, with_vcs: bool = True ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision result = next(vcs_client.log(repository.id, parent=ref, limit=1)) revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision
from dataclasses import dataclass from typing import List, Tuple from zeus.exceptions import UnknownRevision from zeus.models import Repository, Revision from zeus.vcs import vcs_client @dataclass class RevisionResult: sha: str message: str author: str author_date: str committer: str committer_date: str parents: List[str] authors: List[Tuple[str, str]] def identify_revision( repository: Repository, ref: str, with_vcs: bool = True ) -> Revision: """ Attempt to transform a a commit-like reference into a valid revision. """ # try to find it from the database first if len(ref) == 40: revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == ref ).first() if revision: return revision if not with_vcs: raise UnknownRevision try: result = vcs_client.log(repository.id, parent=ref, limit=1)[0] except IndexError: raise UnknownRevision revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first() if not revision: raise UnknownRevision
Fix invalid next() call on api result
Fix invalid next() call on api result
Python
apache-2.0
getsentry/zeus,getsentry/zeus,getsentry/zeus,getsentry/zeus
--- +++ @@ -35,7 +35,11 @@ if not with_vcs: raise UnknownRevision - result = next(vcs_client.log(repository.id, parent=ref, limit=1)) + try: + result = vcs_client.log(repository.id, parent=ref, limit=1)[0] + except IndexError: + raise UnknownRevision + revision = Revision.query.filter( Revision.repository_id == repository.id, Revision.sha == result["sha"] ).first()
4ab14b3de299b58aee94511910d199cd1d1737a5
zou/app/utils/emails.py
zou/app/utils/emails.py
from flask_mail import Message from zou.app import mail def send_email(subject, body, recipient_email, html=None): """ Send an email with given subject and body to given recipient. """ if html is None: html = body message = Message( body=body, html=html, subject=subject, recipients=[recipient_email] ) mail.send(message)
from flask_mail import Message from zou.app import mail, app def send_email(subject, body, recipient_email, html=None): """ Send an email with given subject and body to given recipient. """ if html is None: html = body with app.app_context(): message = Message( body=body, html=html, subject=subject, recipients=[recipient_email] ) mail.send(message)
Fix email sending in production environment
Fix email sending in production environment
Python
agpl-3.0
cgwire/zou
--- +++ @@ -1,6 +1,6 @@ from flask_mail import Message -from zou.app import mail +from zou.app import mail, app def send_email(subject, body, recipient_email, html=None): @@ -9,10 +9,11 @@ """ if html is None: html = body - message = Message( - body=body, - html=html, - subject=subject, - recipients=[recipient_email] - ) - mail.send(message) + with app.app_context(): + message = Message( + body=body, + html=html, + subject=subject, + recipients=[recipient_email] + ) + mail.send(message)
8639f91fba318c4b8c64f7c25885f8fe95e0ebe4
robot/game.py
robot/game.py
import logging import time from enum import Enum from threading import Thread from typing import NewType import _thread from robot.board import Board Zone = NewType('Zone', int) LOGGER = logging.getLogger(__name__) def kill_after_delay(timeout_seconds): """ Interrupts main process after the given delay. """ end_time = time.time() + timeout_seconds def worker(): while time.time() < end_time: remaining = end_time - time.time() time.sleep(max(remaining, 0.01)) LOGGER.info("Timeout %rs expired: Game over!", timeout_seconds) # Interrupt the main thread to kill the user code _thread.interrupt_main() # type: ignore worker_thread = Thread(target=worker, daemon=True) worker_thread.start() return worker_thread class GameMode(Enum): """Possible modes the robot can be in.""" COMPETITION = 'competition' DEVELOPMENT = 'development' class GameState(Board): """A description of the initial game state the robot is operating under.""" @property def zone(self) -> Zone: """ The zone in which the robot starts the match. This is configured by inserting a competition zone USB stick into the robot. :return: zone ID the robot started in (0-3) """ return self._send_and_receive({})['zone'] @property def mode(self) -> GameMode: """ :return: The ``GameMode`` that the robot is currently in. """ value = self._send_and_receive({})['mode'] return GameMode(value)
import logging import signal from enum import Enum from typing import NewType from robot.board import Board Zone = NewType('Zone', int) LOGGER = logging.getLogger(__name__) def timeout_handler(signum, stack): """ Handle the `SIGALRM` to kill the current process. """ raise SystemExit("Timeout expired: Game Over!") def kill_after_delay(timeout_seconds): """ Interrupts main process after the given delay. """ signal.signal(signal.SIGALRM, timeout_handler) signal.alarm(timeout_seconds) class GameMode(Enum): """Possible modes the robot can be in.""" COMPETITION = 'competition' DEVELOPMENT = 'development' class GameState(Board): """A description of the initial game state the robot is operating under.""" @property def zone(self) -> Zone: """ The zone in which the robot starts the match. This is configured by inserting a competition zone USB stick into the robot. :return: zone ID the robot started in (0-3) """ return self._send_and_receive({})['zone'] @property def mode(self) -> GameMode: """ :return: The ``GameMode`` that the robot is currently in. """ value = self._send_and_receive({})['mode'] return GameMode(value)
Replace thread killing with SIGALRM
Replace thread killing with SIGALRM
Python
mit
sourcebots/robot-api,sourcebots/robot-api
--- +++ @@ -1,10 +1,8 @@ import logging -import time +import signal from enum import Enum -from threading import Thread from typing import NewType -import _thread from robot.board import Board Zone = NewType('Zone', int) @@ -12,26 +10,20 @@ LOGGER = logging.getLogger(__name__) +def timeout_handler(signum, stack): + """ + Handle the `SIGALRM` to kill the current process. + """ + raise SystemExit("Timeout expired: Game Over!") + + def kill_after_delay(timeout_seconds): """ Interrupts main process after the given delay. """ - end_time = time.time() + timeout_seconds - - def worker(): - while time.time() < end_time: - remaining = end_time - time.time() - time.sleep(max(remaining, 0.01)) - - LOGGER.info("Timeout %rs expired: Game over!", timeout_seconds) - - # Interrupt the main thread to kill the user code - _thread.interrupt_main() # type: ignore - - worker_thread = Thread(target=worker, daemon=True) - worker_thread.start() - return worker_thread + signal.signal(signal.SIGALRM, timeout_handler) + signal.alarm(timeout_seconds) class GameMode(Enum):
5c88f210644cbe59cf3b3a71345a3fc64dfc542a
spiff/api/plugins.py
spiff/api/plugins.py
from django.conf import settings import importlib import inspect def find_api_classes(module, superclass, test=lambda x: True): for app in map(lambda x:'%s.%s'%(x, module), settings.INSTALLED_APPS): try: appAPI = importlib.import_module(app) except ImportError, e: continue for name, cls in inspect.getmembers(appAPI): if inspect.isclass(cls) and issubclass(cls, superclass) and not cls is superclass and test(cls): yield cls
from django.conf import settings import importlib import inspect def find_api_classes(*args, **kwargs): for app, cls in find_api_implementations(*args, **kwargs): yield cls def find_api_implementations(module, superclass, test=lambda x: True): for app in map(lambda x:'%s.%s'%(x, module), settings.INSTALLED_APPS): try: appAPI = importlib.import_module(app) except ImportError, e: continue for name, cls in inspect.getmembers(appAPI): if inspect.isclass(cls) and issubclass(cls, superclass) and not cls is superclass and test(cls): yield (app, cls)
Add a method to also easily find what app an api object was found in
Add a method to also easily find what app an api object was found in
Python
agpl-3.0
SYNHAK/spiff,SYNHAK/spiff,SYNHAK/spiff
--- +++ @@ -2,7 +2,11 @@ import importlib import inspect -def find_api_classes(module, superclass, test=lambda x: True): +def find_api_classes(*args, **kwargs): + for app, cls in find_api_implementations(*args, **kwargs): + yield cls + +def find_api_implementations(module, superclass, test=lambda x: True): for app in map(lambda x:'%s.%s'%(x, module), settings.INSTALLED_APPS): try: appAPI = importlib.import_module(app) @@ -10,5 +14,4 @@ continue for name, cls in inspect.getmembers(appAPI): if inspect.isclass(cls) and issubclass(cls, superclass) and not cls is superclass and test(cls): - yield cls - + yield (app, cls)
d36ac9a113608aadbda79c724f6aa6f6da5ec0bd
cellcounter/mixins.py
cellcounter/mixins.py
import simplejson as json from django.http import HttpResponse class JSONResponseMixin(object): """ A Mixin that renders context as a JSON response """ def render_to_response(self, context): """ Returns a JSON response containing 'context' as payload """ return self.get_json_response(self.convert_context_to_json(context)) def get_json_response(self, content, **httpresponse_kwargs): """ Construct an `HttpResponse` object. """ response = HttpResponse(content, content_type='application/json', **httpresponse_kwargs) return response def convert_context_to_json(self, context): """ Convert the context dictionary into a JSON object """ return json.dumps(context, indent=4)
import json from django.http import HttpResponse class JSONResponseMixin(object): """ A Mixin that renders context as a JSON response """ def render_to_response(self, context): """ Returns a JSON response containing 'context' as payload """ return self.get_json_response(self.convert_context_to_json(context)) def get_json_response(self, content, **httpresponse_kwargs): """ Construct an `HttpResponse` object. """ response = HttpResponse(content, content_type='application/json', **httpresponse_kwargs) return response def convert_context_to_json(self, context): """ Convert the context dictionary into a JSON object """ return json.dumps(context, indent=4)
Use json rather than simplejson
Use json rather than simplejson
Python
mit
haematologic/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,cellcounter/cellcounter,haematologic/cellcounter,haematologic/cellcounter
--- +++ @@ -1,4 +1,4 @@ -import simplejson as json +import json from django.http import HttpResponse
85e10e4c4eaf46ed89bc4b148b9c483df79cf410
test/test_fields.py
test/test_fields.py
import numpy as np import pyfds as fds def test_dimension(): dim = fds.Dimension(3, 0.1) assert np.allclose(dim.vector, np.asarray([0, 0.1, 0.2])) assert dim.get_index(0.1) == 1
import numpy as np import pyfds as fds def test_dimension(): dim = fds.Dimension(3, 0.1) assert np.allclose(dim.vector, np.asarray([0, 0.1, 0.2])) assert dim.get_index(0.1) == 1 def test_field_component_boundary_1(): fc = fds.FieldComponent(100) fc.values = np.random.rand(100) fc.boundaries = [fds.Boundary(fds.LineRegion([5, 6, 7], [0, 0.2], 'test boundary'))] fc.boundaries[0].value = 23 fc.apply_bounds() assert np.allclose(fc.values[[5, 6, 7]], [23, 23, 23]) def test_field_component_boundary_2(): fc = fds.FieldComponent(100) fc.values = np.ones(100) fc.boundaries = [fds.Boundary(fds.LineRegion([5, 6, 7], [0, 0.2], 'test boundary'))] fc.boundaries[0].value = [23, 42, 23] fc.boundaries[0].additive = True fc.apply_bounds() assert np.allclose(fc.values[[5, 6, 7]], [24, 43, 24]) def test_field_component_output(): fc = fds.FieldComponent(100) fc.outputs = [fds.Output(fds.LineRegion([0, 1, 2], [0, 0.2], 'test output'))] fc.write_outputs() fc.write_outputs() assert np.allclose(fc.outputs[0].signals, [[0, 0], [0, 0], [0, 0]])
Add test cases for FieldComponent class.
Add test cases for FieldComponent class.
Python
bsd-3-clause
emtpb/pyfds
--- +++ @@ -6,3 +6,30 @@ dim = fds.Dimension(3, 0.1) assert np.allclose(dim.vector, np.asarray([0, 0.1, 0.2])) assert dim.get_index(0.1) == 1 + + +def test_field_component_boundary_1(): + fc = fds.FieldComponent(100) + fc.values = np.random.rand(100) + fc.boundaries = [fds.Boundary(fds.LineRegion([5, 6, 7], [0, 0.2], 'test boundary'))] + fc.boundaries[0].value = 23 + fc.apply_bounds() + assert np.allclose(fc.values[[5, 6, 7]], [23, 23, 23]) + + +def test_field_component_boundary_2(): + fc = fds.FieldComponent(100) + fc.values = np.ones(100) + fc.boundaries = [fds.Boundary(fds.LineRegion([5, 6, 7], [0, 0.2], 'test boundary'))] + fc.boundaries[0].value = [23, 42, 23] + fc.boundaries[0].additive = True + fc.apply_bounds() + assert np.allclose(fc.values[[5, 6, 7]], [24, 43, 24]) + + +def test_field_component_output(): + fc = fds.FieldComponent(100) + fc.outputs = [fds.Output(fds.LineRegion([0, 1, 2], [0, 0.2], 'test output'))] + fc.write_outputs() + fc.write_outputs() + assert np.allclose(fc.outputs[0].signals, [[0, 0], [0, 0], [0, 0]])
32d23eb7764178cedcf6b648f959fbf49d7ff657
app/models.py
app/models.py
from . import db class Essay(db.Model): __tablename__ = 'essays' id = db.Column(db.Integer, primary_key=True) text = db.Column(db.Text) time = db.Column(db.DateTime(True)) score = db.Column(db.Float) spell_errors = db.Column(db.Text) grammar_errors = db.Column(db.Text) coherence = db.Column(db.Text) def __repr__(self): return '<Essay {0}>: {1}. Created at:{2}'.format(self.text, self.score, self.time)
from . import db class Essay(db.Model): __tablename__ = 'essays' id = db.Column(db.Integer, primary_key=True) text = db.Column(db.Text) time = db.Column(db.DateTime(True)) score = db.Column(db.Float) spell_errors = db.Column(db.Text) grammar_errors = db.Column(db.Text) coherence = db.Column(db.Text) def __repr__(self): return u'<Essay {0}>: {1}. Created at:{2}'.format(self.id, self.score, self.time)
Modify repr of model Essay
Modify repr of model Essay
Python
apache-2.0
kigawas/essai,kigawas/essai
--- +++ @@ -12,5 +12,5 @@ coherence = db.Column(db.Text) def __repr__(self): - return '<Essay {0}>: {1}. Created at:{2}'.format(self.text, self.score, + return u'<Essay {0}>: {1}. Created at:{2}'.format(self.id, self.score, self.time)
845192bb91a2421c54a9bbb924e1e09e700aee66
Lib/dialogKit/__init__.py
Lib/dialogKit/__init__.py
""" dialogKit: easy bake dialogs """ # determine the environment try: import FL haveFL = True except ImportError: haveFL = False try: import vanilla haveVanilla = True except ImportError: haveVanilla = False # perform the environment specific import if haveFL: from _dkFL import * if haveVanilla: from _dkVanilla import * else: raise ImportError, 'dialogKit is not available in this environment' numberVersion = (0, 0, "beta", 1) version = "0.0.1b"
""" dialogKit: easy bake dialogs """ # determine the environment haveFL = False haveVanilla = False try: import FL haveFL = True except ImportError: pass if not haveFL: try: import vanilla haveVanilla = True except ImportError: pass # perform the environment specific import if haveFL: from _dkFL import * elif haveVanilla: from _dkVanilla import * else: raise ImportError, 'dialogKit is not available in this environment' numberVersion = (0, 0, "beta", 1) version = "0.0.1b"
Stop trying imports after something has been successfully loaded.
Stop trying imports after something has been successfully loaded.
Python
mit
anthrotype/dialogKit,daltonmaag/dialogKit,typesupply/dialogKit
--- +++ @@ -3,20 +3,23 @@ """ # determine the environment +haveFL = False +haveVanilla = False try: import FL haveFL = True except ImportError: - haveFL = False -try: - import vanilla - haveVanilla = True -except ImportError: - haveVanilla = False + pass +if not haveFL: + try: + import vanilla + haveVanilla = True + except ImportError: + pass # perform the environment specific import if haveFL: from _dkFL import * -if haveVanilla: +elif haveVanilla: from _dkVanilla import * else: raise ImportError, 'dialogKit is not available in this environment'
992e0e2f50418bd87052741f7f1937f8efd052c0
tests/mpath_test.py
tests/mpath_test.py
import unittest import os from utils import create_sparse_tempfile from gi.repository import BlockDev if not BlockDev.is_initialized(): BlockDev.init(None, None) class MpathTestCase(unittest.TestCase): def setUp(self): self.dev_file = create_sparse_tempfile("mpath_test", 1024**3) succ, loop = BlockDev.loop_setup(self.dev_file) if not succ: raise RuntimeError("Failed to setup loop device for testing") self.loop_dev = "/dev/%s" % loop def test_is_mpath_member(self): """Verify that is_mpath_member works as expected""" # just test that some non-mpath is not reported as a multipath member # device and no error is reported self.assertFalse(BlockDev.mpath_is_mpath_member("/dev/loop0")) def tearDown(self): succ = BlockDev.loop_teardown(self.loop_dev) if not succ: os.unlink(self.dev_file) raise RuntimeError("Failed to tear down loop device used for testing") os.unlink(self.dev_file)
import unittest import os from utils import create_sparse_tempfile from gi.repository import BlockDev if not BlockDev.is_initialized(): BlockDev.init(None, None) class MpathTestCase(unittest.TestCase): def setUp(self): self.dev_file = create_sparse_tempfile("mpath_test", 1024**3) succ, loop = BlockDev.loop_setup(self.dev_file) if not succ: raise RuntimeError("Failed to setup loop device for testing") self.loop_dev = "/dev/%s" % loop def tearDown(self): succ = BlockDev.loop_teardown(self.loop_dev) if not succ: os.unlink(self.dev_file) raise RuntimeError("Failed to tear down loop device used for testing") os.unlink(self.dev_file) def test_is_mpath_member(self): """Verify that is_mpath_member works as expected""" # just test that some non-mpath is not reported as a multipath member # device and no error is reported self.assertFalse(BlockDev.mpath_is_mpath_member("/dev/loop0"))
Make the tearDown method of the mpath test case better visible
Make the tearDown method of the mpath test case better visible By moving it to the beginning of the file.
Python
lgpl-2.1
vpodzime/libblockdev,atodorov/libblockdev,vpodzime/libblockdev,vpodzime/libblockdev,snbueno/libblockdev,dashea/libblockdev,atodorov/libblockdev,snbueno/libblockdev,rhinstaller/libblockdev,atodorov/libblockdev,rhinstaller/libblockdev,dashea/libblockdev,rhinstaller/libblockdev
--- +++ @@ -14,13 +14,6 @@ raise RuntimeError("Failed to setup loop device for testing") self.loop_dev = "/dev/%s" % loop - def test_is_mpath_member(self): - """Verify that is_mpath_member works as expected""" - - # just test that some non-mpath is not reported as a multipath member - # device and no error is reported - self.assertFalse(BlockDev.mpath_is_mpath_member("/dev/loop0")) - def tearDown(self): succ = BlockDev.loop_teardown(self.loop_dev) if not succ: @@ -28,3 +21,10 @@ raise RuntimeError("Failed to tear down loop device used for testing") os.unlink(self.dev_file) + + def test_is_mpath_member(self): + """Verify that is_mpath_member works as expected""" + + # just test that some non-mpath is not reported as a multipath member + # device and no error is reported + self.assertFalse(BlockDev.mpath_is_mpath_member("/dev/loop0"))
437c8b59148ccb31ac7480ab6c9e9784e2dd6295
js2xml/__init__.py
js2xml/__init__.py
import lxml.etree from slimit.parser import Parser from js2xml.xmlvisitor import XmlVisitor _parser = Parser() _visitor = XmlVisitor() def parse(text, debug=False): tree = _parser.parse(text, debug=debug) xml = _visitor.visit(tree) return xml
Add js2xml.parse() method that wraps the slimit visitor/xml-builder
Add js2xml.parse() method that wraps the slimit visitor/xml-builder
Python
mit
redapple/js2xml,redapple/js2xml,redapple/js2xml,redapple/js2xml
--- +++ @@ -1 +1,11 @@ +import lxml.etree +from slimit.parser import Parser +from js2xml.xmlvisitor import XmlVisitor +_parser = Parser() +_visitor = XmlVisitor() + +def parse(text, debug=False): + tree = _parser.parse(text, debug=debug) + xml = _visitor.visit(tree) + return xml
290f0beb0103ee8f8d3d59bf2fabc227ed743d30
lib/smisk/mvc/model.py
lib/smisk/mvc/model.py
# encoding: utf-8 '''Model in MVC :requires: `elixir <http://elixir.ematia.de/>`__ ''' # Ignore the SA string type depr warning from sqlalchemy.exceptions import SADeprecationWarning from warnings import filterwarnings filterwarnings('ignore', 'Using String type with no length for CREATE TABLE', SADeprecationWarning) # Import Elixir & SQLAlchemy from elixir import * from sqlalchemy import func # Disable autosetup by recommendation from Jason R. Coombs: # http://groups.google.com/group/sqlelixir/msg/ed698d986bfeefdb options_defaults['autosetup'] = False
# encoding: utf-8 '''Model in MVC :requires: `elixir <http://elixir.ematia.de/>`__ ''' # Ignore the SA string type depr warning from sqlalchemy.exceptions import SADeprecationWarning from warnings import filterwarnings filterwarnings('ignore', 'Using String type with no length for CREATE TABLE', SADeprecationWarning) # Import Elixir & SQLAlchemy from elixir import * from sqlalchemy import func # Disable autosetup by recommendation from Jason R. Coombs: # http://groups.google.com/group/sqlelixir/msg/ed698d986bfeefdb options_defaults['autosetup'] = False # Control wheretere to include module name or not in table names. # If True, project.fruits.Apple -> table apples. # If False, project.fruits.Apple -> table project_fruits_apples. options_defaults['shortnames'] = True
Set sqlalchemy option shortnames to True by default.
Set sqlalchemy option shortnames to True by default.
Python
mit
rsms/smisk,rsms/smisk,rsms/smisk
--- +++ @@ -17,3 +17,8 @@ # Disable autosetup by recommendation from Jason R. Coombs: # http://groups.google.com/group/sqlelixir/msg/ed698d986bfeefdb options_defaults['autosetup'] = False + +# Control wheretere to include module name or not in table names. +# If True, project.fruits.Apple -> table apples. +# If False, project.fruits.Apple -> table project_fruits_apples. +options_defaults['shortnames'] = True
80b8ef8b227baa1f4af842716ebfb83dcabf9703
tests/scoring_engine/web/views/test_auth.py
tests/scoring_engine/web/views/test_auth.py
from tests.scoring_engine.web.web_test import WebTest class TestAuth(WebTest): def test_login_page_auth_required(self): resp = self.client.get('/login') assert resp.status_code == 200 def test_unauthorized(self): resp = self.client.get('/unauthorized') assert resp.status_code == 200 def test_auth_required_logout(self): self.verify_auth_required('/logout') def test_login_logout(self): user = self.create_default_user() assert user.authenticated is False self.auth_and_get_path('/') assert user.authenticated is True logout_resp = self.client.get('/logout') assert user.authenticated is False assert logout_resp.status_code == 302 self.verify_auth_required('/services')
from tests.scoring_engine.web.web_test import WebTest class TestAuth(WebTest): def test_login_page_auth_required(self): resp = self.client.get('/login') assert resp.status_code == 200 def test_unauthorized(self): resp = self.client.get('/unauthorized') assert resp.status_code == 200 def test_auth_required_logout(self): self.verify_auth_required('/logout') def test_login_logout(self): user = self.create_default_user() assert user.authenticated is False self.auth_and_get_path('/') assert user.authenticated is True logout_resp = self.client.get('/logout') assert user.authenticated is False assert logout_resp.status_code == 302 self.verify_auth_required('/services') def test_wrong_password_login(self): user = self.create_default_user() user.username = 'RandomName' self.session.add(user) self.session.commit() self.auth_and_get_path('/') assert user.authenticated is False
Add test for incorrect password auth view
Add test for incorrect password auth view
Python
mit
pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine
--- +++ @@ -23,3 +23,11 @@ assert user.authenticated is False assert logout_resp.status_code == 302 self.verify_auth_required('/services') + + def test_wrong_password_login(self): + user = self.create_default_user() + user.username = 'RandomName' + self.session.add(user) + self.session.commit() + self.auth_and_get_path('/') + assert user.authenticated is False
a23774fe4646db1bf4b25cf67d855dcc4cc4c8f7
celery/loaders/default.py
celery/loaders/default.py
from celery.loaders.base import BaseLoader DEFAULT_SETTINGS = { "DEBUG": False, "DATABASE_ENGINE": "sqlite3", "DATABASE_NAME": "celery.sqlite", "INSTALLED_APPS": ("celery", ), } def wanted_module_item(item): is_private = item.startswith("_") return not is_private class Loader(BaseLoader): """The default loader. See the FAQ for example usage. """ def read_configuration(self): """Read configuration from ``celeryconf.py`` and configure celery and Django so it can be used by regular Python.""" config = dict(DEFAULT_SETTINGS) import celeryconfig usercfg = dict((key, getattr(celeryconfig, key)) for key in dir(celeryconfig) if wanted_module_item(key)) config.update(usercfg) from django.conf import settings if not settings.configured: settings.configure() for config_key, config_value in usercfg.items(): setattr(settings, config_key, config_value) return settings def on_worker_init(self): """Imports modules at worker init so tasks can be registered and used by the worked. The list of modules to import is taken from the ``CELERY_IMPORTS`` setting in ``celeryconf.py``. """ imports = getattr(self.conf, "CELERY_IMPORTS", []) for module in imports: __import__(module, [], [], [''])
import os from celery.loaders.base import BaseLoader DEFAULT_CONFIG_MODULE = "celeryconfig" DEFAULT_SETTINGS = { "DEBUG": False, "DATABASE_ENGINE": "sqlite3", "DATABASE_NAME": "celery.sqlite", "INSTALLED_APPS": ("celery", ), } def wanted_module_item(item): is_private = item.startswith("_") return not is_private class Loader(BaseLoader): """The default loader. See the FAQ for example usage. """ def read_configuration(self): """Read configuration from ``celeryconfig.py`` and configure celery and Django so it can be used by regular Python.""" config = dict(DEFAULT_SETTINGS) configname = os.environ.get("CELERY_CONFIG_MODULE", DEFAULT_CONFIG_MODULE) celeryconfig = __import__(configname, {}, {}, ['']) import celeryconfig usercfg = dict((key, getattr(celeryconfig, key)) for key in dir(celeryconfig) if wanted_module_item(key)) config.update(usercfg) from django.conf import settings if not settings.configured: settings.configure() for config_key, config_value in usercfg.items(): setattr(settings, config_key, config_value) return settings def on_worker_init(self): """Imports modules at worker init so tasks can be registered and used by the worked. The list of modules to import is taken from the ``CELERY_IMPORTS`` setting in ``celeryconf.py``. """ imports = getattr(self.conf, "CELERY_IMPORTS", []) for module in imports: __import__(module, [], [], [''])
Add possibility to set celeryconfig module with ENV["CELERY_CONFIG_MODULE"] + always add celery to INSTALLED_APPS
Add possibility to set celeryconfig module with ENV["CELERY_CONFIG_MODULE"] + always add celery to INSTALLED_APPS
Python
bsd-3-clause
ask/celery,frac/celery,mitsuhiko/celery,ask/celery,cbrepo/celery,frac/celery,mitsuhiko/celery,WoLpH/celery,WoLpH/celery,cbrepo/celery
--- +++ @@ -1,4 +1,7 @@ +import os from celery.loaders.base import BaseLoader + +DEFAULT_CONFIG_MODULE = "celeryconfig" DEFAULT_SETTINGS = { "DEBUG": False, @@ -21,9 +24,12 @@ """ def read_configuration(self): - """Read configuration from ``celeryconf.py`` and configure + """Read configuration from ``celeryconfig.py`` and configure celery and Django so it can be used by regular Python.""" config = dict(DEFAULT_SETTINGS) + configname = os.environ.get("CELERY_CONFIG_MODULE", + DEFAULT_CONFIG_MODULE) + celeryconfig = __import__(configname, {}, {}, ['']) import celeryconfig usercfg = dict((key, getattr(celeryconfig, key)) for key in dir(celeryconfig)
349bb1ce2c15239ae3f9c066ed774b20369b9c0d
src/ggrc/settings/app_engine.py
src/ggrc/settings/app_engine.py
# Copyright (C) 2013 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: dan@reciprocitylabs.com # Maintained By: dan@reciprocitylabs.com APP_ENGINE = True ENABLE_JASMINE = False LOGIN_MANAGER = 'ggrc.login.appengine' FULLTEXT_INDEXER = 'ggrc.fulltext.mysql.MysqlIndexer' # Cannot access filesystem on AppEngine or when using SDK AUTOBUILD_ASSETS = False SQLALCHEMY_RECORD_QUERIES = True MEMCACHE_MECHANISM = True
# Copyright (C) 2013 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: dan@reciprocitylabs.com # Maintained By: dan@reciprocitylabs.com APP_ENGINE = True ENABLE_JASMINE = False LOGIN_MANAGER = 'ggrc.login.appengine' FULLTEXT_INDEXER = 'ggrc.fulltext.mysql.MysqlIndexer' # Cannot access filesystem on AppEngine or when using SDK AUTOBUILD_ASSETS = False SQLALCHEMY_RECORD_QUERIES = True MEMCACHE_MECHANISM = True CALENDAR_MECHANISM = True
Enable Calendar integration on App Engine deployments
Enable Calendar integration on App Engine deployments
Python
apache-2.0
NejcZupec/ggrc-core,uskudnik/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,edofic/ggrc-core,kr41/ggrc-core,hasanalom/ggrc-core,andrei-karalionak/ggrc-core,hyperNURb/ggrc-core,edofic/ggrc-core,selahssea/ggrc-core,vladan-m/ggrc-core,jmakov/ggrc-core,prasannav7/ggrc-core,vladan-m/ggrc-core,VinnieJohns/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,andrei-karalionak/ggrc-core,kr41/ggrc-core,plamut/ggrc-core,andrei-karalionak/ggrc-core,selahssea/ggrc-core,prasannav7/ggrc-core,plamut/ggrc-core,selahssea/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,prasannav7/ggrc-core,uskudnik/ggrc-core,VinnieJohns/ggrc-core,hyperNURb/ggrc-core,josthkko/ggrc-core,josthkko/ggrc-core,hasanalom/ggrc-core,kr41/ggrc-core,jmakov/ggrc-core,uskudnik/ggrc-core,josthkko/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core,AleksNeStu/ggrc-core,AleksNeStu/ggrc-core,hasanalom/ggrc-core,j0gurt/ggrc-core,NejcZupec/ggrc-core,plamut/ggrc-core,AleksNeStu/ggrc-core,hasanalom/ggrc-core,uskudnik/ggrc-core,edofic/ggrc-core,uskudnik/ggrc-core,josthkko/ggrc-core,jmakov/ggrc-core,plamut/ggrc-core,NejcZupec/ggrc-core,edofic/ggrc-core,jmakov/ggrc-core,hasanalom/ggrc-core,jmakov/ggrc-core,hyperNURb/ggrc-core,vladan-m/ggrc-core,hyperNURb/ggrc-core,VinnieJohns/ggrc-core,hyperNURb/ggrc-core,vladan-m/ggrc-core,vladan-m/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core
--- +++ @@ -11,3 +11,4 @@ AUTOBUILD_ASSETS = False SQLALCHEMY_RECORD_QUERIES = True MEMCACHE_MECHANISM = True +CALENDAR_MECHANISM = True
5a84249b7e96d9d2f82ee1b27a33b7978d63b16e
src/urls.py
src/urls.py
# -*- coding: utf-8 -*- # urls.py used as base for developing wirecloud. try: from django.conf.urls import patterns, include, url except ImportError: # pragma: no cover # for Django version less than 1.4 from django.conf.urls.defaults import patterns, include, url from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns import wirecloud.platform.urls admin.autodiscover() urlpatterns = patterns('', # Showcase (r'^showcase/', include('wirecloud.platform.widget.showcase_urls')), # Catalogue (r'^catalogue', include('wirecloud.catalogue.urls')), # Proxy (r'^proxy', include('wirecloud.proxy.urls')), # Login/logout url(r'^login/?$', 'django.contrib.auth.views.login', name="login"), url(r'^logout/?$', 'wirecloud.commons.authentication.logout', name="logout"), url(r'^admin/logout/?$', 'wirecloud.commons.authentication.logout'), # Admin interface (r'^admin/', include(admin.site.urls)), ) urlpatterns += wirecloud.platform.urls.urlpatterns urlpatterns += staticfiles_urlpatterns() handler404 = "django.views.defaults.page_not_found" handler500 = "wirecloud.commons.views.server_error"
# -*- coding: utf-8 -*- # urls.py used as base for developing wirecloud. from django.conf.urls import patterns, include, url from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns import wirecloud.platform.urls admin.autodiscover() urlpatterns = patterns('', # Showcase (r'^showcase/', include('wirecloud.platform.widget.showcase_urls')), # Catalogue (r'^catalogue', include('wirecloud.catalogue.urls')), # Proxy (r'^proxy', include('wirecloud.proxy.urls')), # Login/logout url(r'^login/?$', 'django.contrib.auth.views.login', name="login"), url(r'^logout/?$', 'wirecloud.commons.authentication.logout', name="logout"), url(r'^admin/logout/?$', 'wirecloud.commons.authentication.logout'), # Admin interface (r'^admin/', include(admin.site.urls)), ) urlpatterns += wirecloud.platform.urls.urlpatterns urlpatterns += staticfiles_urlpatterns() handler404 = "django.views.defaults.page_not_found" handler500 = "wirecloud.commons.views.server_error"
Remove django < 1.4 code
Remove django < 1.4 code
Python
agpl-3.0
rockneurotiko/wirecloud,jpajuelo/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud
--- +++ @@ -1,11 +1,7 @@ # -*- coding: utf-8 -*- # urls.py used as base for developing wirecloud. -try: - from django.conf.urls import patterns, include, url -except ImportError: # pragma: no cover - # for Django version less than 1.4 - from django.conf.urls.defaults import patterns, include, url +from django.conf.urls import patterns, include, url from django.contrib import admin from django.contrib.staticfiles.urls import staticfiles_urlpatterns
8ec8929595bd7c4d9b794fe016da64532e517a53
producers/producers.py
producers/producers.py
class Producer(object): """ Base class for producers. __init__ must be called by inheriting classes. Inheriting classes must implement: - ``_run`` - to run the producer - ``configure(jvm, *options)`` - to configure itself with the given jvm and options (must set configured to True if fully configured) """ DEPENDENCIES = [] def __init__(self): self._pre_hooks = [] self._post_hooks = [] self.configured = False def add_pre_hooks(self, *hooks): self._pre_hooks.extend(hooks) def add_post_hooks(self, *hooks): self._post_hooks.extend(hooks) def _run(): raise NotImplementedError("_run must be implemented by actual producers") def configure(jvm, **options): """ Configure producer with :param:`jvm` and :param:`options`. :param jvm: Instance of an JVM-Class :param options: keywords that are understood by ``JVM.get_commandline()`` """ raise NotImplementedError("configure must be implemented by actual producers") def run(): """ Run Producer """ for hook in self._pre_hooks: hook() self.run_() for hook in self._post_hooks: hook() class Dacapo(Producer): #TODO pass class Tamiflex(Producer): #TODO pass class HProf(Producer): #TODO pass
class Producer(object): """ Base class for producers. __init__ must be called by inheriting classes. Inheriting classes must implement: - ``_run`` to run the producer, after running `out` attribute has to be set to path to produced output - ``configure(jvm, *options)`` to configure itself with the given jvm and options (must set configured to True if fully configured) - ``is_runable`` to state if producer can be run in current state """ DEPENDENCIES = [] def __init__(self): self._pre_hooks = [] self._post_hooks = [] self.configured = False def add_pre_hooks(self, *hooks): self._pre_hooks.extend(hooks) def add_post_hooks(self, *hooks): self._post_hooks.extend(hooks) def _run(): raise NotImplementedError("_run must be implemented by actual producers") def is_runable(): """ :returns: if producer is runable :rtype: bool """ raise NotImplementedError("runable must be implemented by actual producers") def configure(jvm, **options): """ Configure producer with :param:`jvm` and :param:`options`. :param jvm: Instance of an JVM-Class :param options: keywords that are understood by ``JVM.get_commandline()`` """ raise NotImplementedError("configure must be implemented by actual producers") def run(): """ Run Producer """ for hook in self._pre_hooks: hook() self.run_() for hook in self._post_hooks: hook() class Dacapo(Producer): #TODO pass class Tamiflex(Producer): #TODO pass class HProf(Producer): #TODO pass
Add `is_runable` to Producer base class.
Add `is_runable` to Producer base class. Signed-off-by: Michael Markert <5eb998b7ac86da375651a4cd767b88c9dad25896@googlemail.com>
Python
mit
fhirschmann/penchy,fhirschmann/penchy
--- +++ @@ -3,9 +3,11 @@ Base class for producers. __init__ must be called by inheriting classes. Inheriting classes must implement: - - ``_run`` - to run the producer - - ``configure(jvm, *options)`` - to configure itself with the given jvm + - ``_run`` to run the producer, after running `out` attribute has to be + set to path to produced output + - ``configure(jvm, *options)`` to configure itself with the given jvm and options (must set configured to True if fully configured) + - ``is_runable`` to state if producer can be run in current state """ DEPENDENCIES = [] def __init__(self): @@ -21,6 +23,14 @@ def _run(): raise NotImplementedError("_run must be implemented by actual producers") + + def is_runable(): + """ + :returns: if producer is runable + :rtype: bool + """ + raise NotImplementedError("runable must be implemented by actual producers") + def configure(jvm, **options): """
b704a92c919d7fa950a65ee0c569864c4549331f
glue/core/tests/util.py
glue/core/tests/util.py
from __future__ import absolute_import, division, print_function import tempfile from contextlib import contextmanager import os import zlib from mock import MagicMock from ... import core from ...core.application_base import Application @contextmanager def make_file(contents, suffix, decompress=False): """Context manager to write data to a temporary file, and delete on exit :param contents: Data to write. string :param suffix: File suffix. string """ if decompress: contents = zlib.decompress(contents) try: _, fname = tempfile.mkstemp(suffix=suffix) with open(fname, 'wb') as outfile: outfile.write(contents) yield fname finally: os.unlink(fname) @contextmanager def simple_catalog(): """Context manager to create a temporary data file :param suffix: File suffix. string """ with make_file(b'#a, b\n1, 2\n3, 4', '.csv') as result: yield result def simple_session(): collect = core.data_collection.DataCollection() hub = core.hub.Hub() result = core.Session(data_collection=collect, hub=hub, application=MagicMock(Application), command_stack=core.CommandStack()) result.command_stack.session = result return result
from __future__ import absolute_import, division, print_function import tempfile from contextlib import contextmanager import os import zlib from mock import MagicMock from ... import core from ...core.application_base import Application @contextmanager def make_file(contents, suffix, decompress=False): """Context manager to write data to a temporary file, and delete on exit :param contents: Data to write. string :param suffix: File suffix. string """ if decompress: contents = zlib.decompress(contents) try: _, fname = tempfile.mkstemp(suffix=suffix) with open(fname, 'wb') as outfile: outfile.write(contents) yield fname finally: try: os.unlink(fname) except WindowsError: # on Windows the unlink can fail pass @contextmanager def simple_catalog(): """Context manager to create a temporary data file :param suffix: File suffix. string """ with make_file(b'#a, b\n1, 2\n3, 4', '.csv') as result: yield result def simple_session(): collect = core.data_collection.DataCollection() hub = core.hub.Hub() result = core.Session(data_collection=collect, hub=hub, application=MagicMock(Application), command_stack=core.CommandStack()) result.command_stack.session = result return result
Add workaround for failing unlink on Windows
Add workaround for failing unlink on Windows
Python
bsd-3-clause
saimn/glue,stscieisenhamer/glue,JudoWill/glue,saimn/glue,JudoWill/glue,stscieisenhamer/glue
--- +++ @@ -28,7 +28,10 @@ outfile.write(contents) yield fname finally: - os.unlink(fname) + try: + os.unlink(fname) + except WindowsError: # on Windows the unlink can fail + pass @contextmanager
17fe6d36a34218e74b53e9617212f0e67b05297d
pysteps/io/__init__.py
pysteps/io/__init__.py
from .interface import get_method from .archive import * from .importers import * from .readers import *
from .interface import get_method from .archive import * from .exporters import * from .importers import * from .readers import *
Add missing import of the exporters module
Add missing import of the exporters module
Python
bsd-3-clause
pySTEPS/pysteps
--- +++ @@ -1,4 +1,5 @@ from .interface import get_method from .archive import * +from .exporters import * from .importers import * from .readers import *
29b7a69a39ac66ebd8f61c6c9c65e7e60b40b4a0
numpy/_array_api/_types.py
numpy/_array_api/_types.py
""" This file defines the types for type annotations. These names aren't part of the module namespace, but they are used in the annotations in the function signatures. The functions in the module are only valid for inputs that match the given type annotations. """ __all__ = ['Array', 'Device', 'Dtype', 'SupportsDLPack', 'SupportsBufferProtocol', 'PyCapsule'] from typing import Any, Sequence, Type, Union from . import (Array, int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64) Array = ndarray Device = TypeVar('device') Dtype = Literal[int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64] SupportsDLPack = TypeVar('SupportsDLPack') SupportsBufferProtocol = TypeVar('SupportsBufferProtocol') PyCapsule = TypeVar('PyCapsule')
""" This file defines the types for type annotations. These names aren't part of the module namespace, but they are used in the annotations in the function signatures. The functions in the module are only valid for inputs that match the given type annotations. """ __all__ = ['Array', 'Device', 'Dtype', 'SupportsDLPack', 'SupportsBufferProtocol', 'PyCapsule'] from typing import Any, Sequence, Type, Union from . import (Array, int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64) # This should really be recursive, but that isn't supported yet. See the # similar comment in numpy/typing/_array_like.py NestedSequence = Sequence[Sequence[Any]] Device = Any Dtype = Type[Union[[int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64]]] SupportsDLPack = Any SupportsBufferProtocol = Any PyCapsule = Any
Use better type definitions for the array API custom types
Use better type definitions for the array API custom types
Python
bsd-3-clause
anntzer/numpy,simongibbons/numpy,jakirkham/numpy,rgommers/numpy,pdebuyl/numpy,endolith/numpy,simongibbons/numpy,mhvk/numpy,pdebuyl/numpy,charris/numpy,rgommers/numpy,jakirkham/numpy,mattip/numpy,mhvk/numpy,mattip/numpy,rgommers/numpy,charris/numpy,numpy/numpy,simongibbons/numpy,endolith/numpy,anntzer/numpy,anntzer/numpy,pdebuyl/numpy,mhvk/numpy,numpy/numpy,jakirkham/numpy,anntzer/numpy,simongibbons/numpy,seberg/numpy,seberg/numpy,simongibbons/numpy,jakirkham/numpy,seberg/numpy,numpy/numpy,mattip/numpy,pdebuyl/numpy,charris/numpy,mattip/numpy,mhvk/numpy,rgommers/numpy,numpy/numpy,seberg/numpy,charris/numpy,jakirkham/numpy,endolith/numpy,mhvk/numpy,endolith/numpy
--- +++ @@ -14,10 +14,13 @@ from . import (Array, int8, int16, int32, int64, uint8, uint16, uint32, uint64, float32, float64) -Array = ndarray -Device = TypeVar('device') -Dtype = Literal[int8, int16, int32, int64, uint8, uint16, - uint32, uint64, float32, float64] -SupportsDLPack = TypeVar('SupportsDLPack') -SupportsBufferProtocol = TypeVar('SupportsBufferProtocol') -PyCapsule = TypeVar('PyCapsule') +# This should really be recursive, but that isn't supported yet. See the +# similar comment in numpy/typing/_array_like.py +NestedSequence = Sequence[Sequence[Any]] + +Device = Any +Dtype = Type[Union[[int8, int16, int32, int64, uint8, uint16, + uint32, uint64, float32, float64]]] +SupportsDLPack = Any +SupportsBufferProtocol = Any +PyCapsule = Any
5277d6d5caf075ce6fbb8d46c558bdc29eb62e19
docs/conf.py
docs/conf.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import setuptools_scm extensions = [ 'sphinx.ext.autodoc', 'rst.linker', ] # General information about the project. project = 'pytest-runner' copyright = '2015,2016 Jason R. Coombs' # The short X.Y version. version = setuptools_scm.get_version(root='..', relative_to=__file__) # The full version, including alpha/beta/rc tags. release = version master_doc = 'index' link_files = { 'CHANGES.rst': dict( using=dict( GH='https://github.com', project=project, ), replace=[ dict( pattern=r"(Issue )?#(?P<issue>\d+)", url='{GH}/jaraco/{project}/issues/{issue}', ), dict( pattern=r"^(?m)((?P<scm_version>v?\d+(\.\d+){1,2}))\n[-=]+\n", with_scm="{text}\n{rev[timestamp]:%d %b %Y}\n", ), ], ), }
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import setuptools_scm extensions = [ 'sphinx.ext.autodoc', 'rst.linker', ] # General information about the project. project = 'pytest-runner' copyright = '2015,2016 Jason R. Coombs' # The short X.Y version. version = setuptools_scm.get_version(root='..', relative_to=__file__) # The full version, including alpha/beta/rc tags. release = version master_doc = 'index' link_files = { 'CHANGES.rst': dict( using=dict( GH='https://github.com', project=project, ), replace=[ dict( pattern=r"(Issue )?#(?P<issue>\d+)", url='{GH}/pytest-dev/{project}/issues/{issue}', ), dict( pattern=r"^(?m)((?P<scm_version>v?\d+(\.\d+){1,2}))\n[-=]+\n", with_scm="{text}\n{rev[timestamp]:%d %b %Y}\n", ), ], ), }
Update org URL for issue linkage
Update org URL for issue linkage
Python
mit
pytest-dev/pytest-runner
--- +++ @@ -28,7 +28,7 @@ replace=[ dict( pattern=r"(Issue )?#(?P<issue>\d+)", - url='{GH}/jaraco/{project}/issues/{issue}', + url='{GH}/pytest-dev/{project}/issues/{issue}', ), dict( pattern=r"^(?m)((?P<scm_version>v?\d+(\.\d+){1,2}))\n[-=]+\n",
56856ac1103ec9f3ba0f2da81832a59e7e773256
doc/ext/nova_autodoc.py
doc/ext/nova_autodoc.py
import os from nova import utils def setup(app): rootdir = os.path.abspath(app.srcdir + '/..') print "**Autodocumenting from %s" % rootdir rv = utils.execute('cd %s && ./generate_autodoc_index.sh' % rootdir) print rv[0]
import gettext import os gettext.install('nova') from nova import utils def setup(app): rootdir = os.path.abspath(app.srcdir + '/..') print "**Autodocumenting from %s" % rootdir rv = utils.execute('cd %s && ./generate_autodoc_index.sh' % rootdir) print rv[0]
Fix doc building endpoint for gettext.
Fix doc building endpoint for gettext.
Python
apache-2.0
blueboxgroup/nova,BeyondTheClouds/nova,russellb/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,SUSE-Cloud/nova,termie/nova-migration-demo,shahar-stratoscale/nova,rajalokan/nova,devendermishrajio/nova_test_latest,aristanetworks/arista-ovs-nova,NoBodyCam/TftpPxeBootBareMetal,virtualopensystems/nova,dims/nova,bclau/nova,gooddata/openstack-nova,anotherjesse/nova,sridevikoushik31/nova,Yuriy-Leonov/nova,zhimin711/nova,tanglei528/nova,sridevikoushik31/openstack,gooddata/openstack-nova,orbitfp7/nova,eneabio/nova,ruslanloman/nova,cyx1231st/nova,ted-gould/nova,dstroppa/openstack-smartos-nova-grizzly,Yusuke1987/openstack_template,anotherjesse/nova,gspilio/nova,zhimin711/nova,paulmathews/nova,josephsuh/extra-specs,KarimAllah/nova,KarimAllah/nova,usc-isi/nova,russellb/nova,leilihh/novaha,yosshy/nova,eonpatapon/nova,badock/nova,adelina-t/nova,openstack/nova,yatinkumbhare/openstack-nova,ted-gould/nova,Yusuke1987/openstack_template,Triv90/Nova,usc-isi/extra-specs,barnsnake351/nova,tudorvio/nova,dawnpower/nova,edulramirez/nova,jianghuaw/nova,cloudbase/nova,maheshp/novatest,luogangyi/bcec-nova,scripnichenko/nova,adelina-t/nova,viggates/nova,double12gzh/nova,CiscoSystems/nova,belmiromoreira/nova,raildo/nova,alexandrucoman/vbox-nova-driver,DirectXMan12/nova-hacking,petrutlucian94/nova,maoy/zknova,savi-dev/nova,dawnpower/nova,TwinkleChawla/nova,mahak/nova,imsplitbit/nova,cernops/nova,cernops/nova,phenoxim/nova,maelnor/nova,silenceli/nova,JioCloud/nova_test_latest,felixma/nova,salv-orlando/MyRepo,berrange/nova,joker946/nova,thomasem/nova,leilihh/novaha,CEG-FYP-OpenStack/scheduler,badock/nova,petrutlucian94/nova_dev,akash1808/nova,alaski/nova,takeshineshiro/nova,orbitfp7/nova,yrobla/nova,termie/pupa,zaina/nova,belmiromoreira/nova,sridevikoushik31/nova,alexandrucoman/vbox-nova-driver,maheshp/novatest,sebrandon1/nova,Tehsmash/nova,houshengbo/nova_vmware_compute_driver,KarimAllah/nova,projectcalico/calico-nova,rickerc/nova_audit,NewpTone/stacklab-nova,barnsnake351/nova,plumgrid/plumgrid-nova,yatinkumbhare/openstack-nova,eneabio/nova,fnordahl/nova,sridevikoushik31/openstack,savi-dev/nova,NewpTone/stacklab-nova,sridevikoushik31/nova,nikesh-mahalka/nova,eharney/nova,maheshp/novatest,psiwczak/openstack,fajoy/nova,rahulunair/nova,DirectXMan12/nova-hacking,yrobla/nova,hanlind/nova,CCI-MOC/nova,noironetworks/nova,devendermishrajio/nova,thomasem/nova,vmturbo/nova,JianyuWang/nova,OpenAcademy-OpenStack/nova-scheduler,klmitch/nova,saleemjaveds/https-github.com-openstack-nova,jeffrey4l/nova,dstroppa/openstack-smartos-nova-grizzly,mikalstill/nova,isyippee/nova,jeffrey4l/nova,ruslanloman/nova,shahar-stratoscale/nova,tanglei528/nova,alaski/nova,isyippee/nova,vmturbo/nova,fajoy/nova,eayunstack/nova,NewpTone/stacklab-nova,mgagne/nova,DirectXMan12/nova-hacking,shail2810/nova,rajalokan/nova,sileht/deb-openstack-nova,blueboxgroup/nova,superstack/nova,Tehsmash/nova,cloudbase/nova-virtualbox,akash1808/nova_test_latest,cernops/nova,gooddata/openstack-nova,zzicewind/nova,alvarolopez/nova,mgagne/nova,eharney/nova,eneabio/nova,Metaswitch/calico-nova,cloudbase/nova,watonyweng/nova,maoy/zknova,whitepages/nova,klmitch/nova,BeyondTheClouds/nova,saleemjaveds/https-github.com-openstack-nova,projectcalico/calico-nova,raildo/nova,tangfeixiong/nova,citrix-openstack-build/nova,Triv90/Nova,j-carpentier/nova,akash1808/nova,josephsuh/extra-specs,Juniper/nova,fajoy/nova,devendermishrajio/nova_test_latest,Juniper/nova,tudorvio/nova,devendermishrajio/nova,usc-isi/extra-specs,usc-isi/nova,BeyondTheClouds/nova,vladikr/nova_drafts,cloudbau/nova,bgxavier/nova,redhat-openstack/nova,termie/pupa,houshengbo/nova_vmware_compute_driver,hanlind/nova,mmnelemane/nova,NoBodyCam/TftpPxeBootBareMetal,TieWei/nova,sridevikoushik31/nova,rrader/nova-docker-plugin,Francis-Liu/animated-broccoli,JianyuWang/nova,rahulunair/nova,TieWei/nova,virtualopensystems/nova,yrobla/nova,gspilio/nova,LoHChina/nova,akash1808/nova_test_latest,kimjaejoong/nova,plumgrid/plumgrid-nova,imsplitbit/nova,apporc/nova,eonpatapon/nova,leilihh/nova,affo/nova,scripnichenko/nova,Stavitsky/nova,bclau/nova,spring-week-topos/nova-week,savi-dev/nova,silenceli/nova,iuliat/nova,OpenAcademy-OpenStack/nova-scheduler,eayunstack/nova,vmturbo/nova,shail2810/nova,TwinkleChawla/nova,LoHChina/nova,varunarya10/nova_test_latest,ntt-sic/nova,openstack/nova,superstack/nova,CCI-MOC/nova,NeCTAR-RC/nova,jianghuaw/nova,rajalokan/nova,Francis-Liu/animated-broccoli,mmnelemane/nova,Yuriy-Leonov/nova,varunarya10/nova_test_latest,josephsuh/extra-specs,tealover/nova,phenoxim/nova,petrutlucian94/nova,sileht/deb-openstack-nova,klmitch/nova,rrader/nova-docker-plugin,edulramirez/nova,double12gzh/nova,paulmathews/nova,iuliat/nova,ntt-sic/nova,felixma/nova,takeshineshiro/nova,mikalstill/nova,usc-isi/extra-specs,vladikr/nova_drafts,russellb/nova,rajalokan/nova,angdraug/nova,salv-orlando/MyRepo,jianghuaw/nova,bigswitch/nova,tianweizhang/nova,sacharya/nova,gooddata/openstack-nova,rickerc/nova_audit,devoid/nova,paulmathews/nova,MountainWei/nova,SUSE-Cloud/nova,qwefi/nova,ewindisch/nova,Triv90/Nova,petrutlucian94/nova_dev,kimjaejoong/nova,superstack/nova,Metaswitch/calico-nova,CiscoSystems/nova,dstroppa/openstack-smartos-nova-grizzly,qwefi/nova,termie/pupa,spring-week-topos/nova-week,sileht/deb-openstack-nova,citrix-openstack-build/nova,berrange/nova,JioCloud/nova_test_latest,houshengbo/nova_vmware_compute_driver,shootstar/novatest,joker946/nova,usc-isi/nova,mahak/nova,sridevikoushik31/openstack,CloudServer/nova,zaina/nova,Juniper/nova,openstack/nova,Brocade-OpenSource/OpenStack-DNRM-Nova,alvarolopez/nova,gspilio/nova,psiwczak/openstack,salv-orlando/MyRepo,affo/nova,leilihh/nova,cloudbau/nova,termie/nova-migration-demo,Stavitsky/nova,mikalstill/nova,JioCloud/nova,klmitch/nova,watonyweng/nova,redhat-openstack/nova,nikesh-mahalka/nova,maoy/zknova,tianweizhang/nova,dims/nova,jianghuaw/nova,zzicewind/nova,apporc/nova,noironetworks/nova,NoBodyCam/TftpPxeBootBareMetal,CEG-FYP-OpenStack/scheduler,tealover/nova,vmturbo/nova,yosshy/nova,cloudbase/nova,mahak/nova,aristanetworks/arista-ovs-nova,sebrandon1/nova,whitepages/nova,ewindisch/nova,shootstar/novatest,angdraug/nova,NeCTAR-RC/nova,sacharya/nova,tangfeixiong/nova,mandeepdhami/nova,anotherjesse/nova,fnordahl/nova,cyx1231st/nova,luogangyi/bcec-nova,viggates/nova,MountainWei/nova,rahulunair/nova,CloudServer/nova,j-carpentier/nova,mandeepdhami/nova,Juniper/nova,maelnor/nova,psiwczak/openstack,bgxavier/nova,cloudbase/nova-virtualbox,bigswitch/nova,sebrandon1/nova,aristanetworks/arista-ovs-nova,termie/nova-migration-demo,devoid/nova,hanlind/nova,JioCloud/nova
--- +++ @@ -1,4 +1,7 @@ +import gettext import os + +gettext.install('nova') from nova import utils
74a2e0825f3029b6d3a3164221d11fbdf551b8d1
demo/demo/widgets/live.py
demo/demo/widgets/live.py
from moksha.api.widgets.live import LiveWidget class HelloWorldWidget(LiveWidget): topic = "helloworld" template = """ <b>Hello World Widget</b> <ul id="data"/> """ onmessage = """ $('<li/>').text(json.msg).prependTo('#data'); """
from moksha.api.widgets.live import LiveWidget class HelloWorldWidget(LiveWidget): topic = "helloworld" template = """ <b>Hello World Widget</b> <form onsubmit="return send_msg()"> <input name="text" id="text"/> </form> <ul id="data"/> <script> function send_msg() { moksha.send_message('helloworld', {'msg': $('#text').val()}); $('#text').val(''); return false; } </script> """ onmessage = """ $('<li/>').text(json.msg).prependTo('#data'); """
Allow people to send messages in our basic HelloWorldWidget demo
Allow people to send messages in our basic HelloWorldWidget demo
Python
apache-2.0
ralphbean/moksha,ralphbean/moksha,lmacken/moksha,mokshaproject/moksha,mokshaproject/moksha,lmacken/moksha,ralphbean/moksha,pombredanne/moksha,mokshaproject/moksha,mokshaproject/moksha,pombredanne/moksha,pombredanne/moksha,pombredanne/moksha,lmacken/moksha
--- +++ @@ -4,7 +4,19 @@ topic = "helloworld" template = """ <b>Hello World Widget</b> + <form onsubmit="return send_msg()"> + <input name="text" id="text"/> + </form> + <ul id="data"/> + + <script> + function send_msg() { + moksha.send_message('helloworld', {'msg': $('#text').val()}); + $('#text').val(''); + return false; + } + </script> """ onmessage = """ $('<li/>').text(json.msg).prependTo('#data');
26581b24dd00c3b0a0928fe0b24ae129c701fb58
jarbas/frontend/tests/test_bundle_dependecies.py
jarbas/frontend/tests/test_bundle_dependecies.py
from django.test import TestCase from webassets.bundle import get_all_bundle_files from jarbas.frontend.assets import elm class TestDependencies(TestCase): def test_dependencies(self): files = set(get_all_bundle_files(elm)) self.assertEqual(9, len(files), files)
from glob import glob from django.test import TestCase from webassets.bundle import get_all_bundle_files from jarbas.frontend.assets import elm class TestDependencies(TestCase): def test_dependencies(self): expected = len(glob('jarbas/frontend/elm/**/*.elm', recursive=True)) files = set(get_all_bundle_files(elm)) self.assertEqual(expected, len(files), files)
Fix test for Elm files lookup
Fix test for Elm files lookup
Python
mit
datasciencebr/jarbas,rogeriochaves/jarbas,Guilhermeslucas/jarbas,marcusrehm/serenata-de-amor,marcusrehm/serenata-de-amor,datasciencebr/jarbas,marcusrehm/serenata-de-amor,Guilhermeslucas/jarbas,datasciencebr/jarbas,Guilhermeslucas/jarbas,rogeriochaves/jarbas,rogeriochaves/jarbas,Guilhermeslucas/jarbas,datasciencebr/jarbas,datasciencebr/serenata-de-amor,rogeriochaves/jarbas,datasciencebr/serenata-de-amor,marcusrehm/serenata-de-amor
--- +++ @@ -1,3 +1,4 @@ +from glob import glob from django.test import TestCase from webassets.bundle import get_all_bundle_files @@ -7,5 +8,6 @@ class TestDependencies(TestCase): def test_dependencies(self): + expected = len(glob('jarbas/frontend/elm/**/*.elm', recursive=True)) files = set(get_all_bundle_files(elm)) - self.assertEqual(9, len(files), files) + self.assertEqual(expected, len(files), files)
cd199c379145c6dcabd66f1771397c82e445c932
test_installation.py
test_installation.py
#!/usr/bin/env python from sys import exit try: import sympy except ImportError: print("SymPy must be installed for the tutorial") if sympy.__version__ != '1.1': print("SymPy 1.1 is required for the tutorial. Note SymPy 1.1 will be released before July 10.") try: import numpy except ImportError: print("NumPy is required for the tutorial") try: import Cython except ImportError: print("Cython is required for the tutorial") try: import scipy except ImportError: print("scipy is required for the tutorial") from sympy.utilities.autowrap import ufuncify from sympy.abc import x from sympy import sin try: f = ufuncify(x, sin(x)) assert f(0) == 0 except: print("sympy.utilities.autowrap.ufuncify does not work") raise
#!/usr/bin/env python from sys import exit try: import sympy except ImportError: print("SymPy must be installed for the tutorial") if sympy.__version__ != '1.1': print("SymPy 1.1 is required for the tutorial. Note SymPy 1.1 will be released before July 10.") try: import numpy except ImportError: print("NumPy is required for the tutorial") try: import Cython except ImportError: print("Cython is required for the tutorial") try: import scipy except ImportError: print("scipy is required for the tutorial") from sympy.utilities.autowrap import ufuncify from sympy.abc import x from sympy import sin try: f = ufuncify(x, sin(x)) assert f(0) == 0 except: print("sympy.utilities.autowrap.ufuncify does not work") raise try: import conda except ImportError: print("conda is needed (either anaconda or miniconda from https://www.continuum.io/downloads)") print("(try rerunning this script under conda if you are using for system's python distribution)") else: major, minor, patch = map(int, conda.__version__.split('.')) if major > 4 or (major == 4 and minor >= 1): pass else: print("please update conda ($ conda update conda), we need conda >= 4.1.0") exit(1) try: import matplotlib except ImportError: print("matplotlib is required for the tutorial") try: import notebook except ImportError: print("notebook (jupyter notebook) is required for the tutorial")
Make test script more extensive (conda, notebook, matplotlib)
Make test script more extensive (conda, notebook, matplotlib)
Python
bsd-3-clause
sympy/scipy-2017-codegen-tutorial,sympy/scipy-2017-codegen-tutorial,sympy/scipy-2017-codegen-tutorial,sympy/scipy-2017-codegen-tutorial,sympy/scipy-2017-codegen-tutorial
--- +++ @@ -35,3 +35,26 @@ except: print("sympy.utilities.autowrap.ufuncify does not work") raise + +try: + import conda +except ImportError: + print("conda is needed (either anaconda or miniconda from https://www.continuum.io/downloads)") + print("(try rerunning this script under conda if you are using for system's python distribution)") +else: + major, minor, patch = map(int, conda.__version__.split('.')) + if major > 4 or (major == 4 and minor >= 1): + pass + else: + print("please update conda ($ conda update conda), we need conda >= 4.1.0") + exit(1) + +try: + import matplotlib +except ImportError: + print("matplotlib is required for the tutorial") + +try: + import notebook +except ImportError: + print("notebook (jupyter notebook) is required for the tutorial")
4a8170079e2b715d40e94f5d407d110a635f8a5d
InvenTree/common/apps.py
InvenTree/common/apps.py
from django.apps import AppConfig from django.db.utils import OperationalError, ProgrammingError, IntegrityError class CommonConfig(AppConfig): name = 'common' def ready(self): """ Will be called when the Common app is first loaded """ self.add_instance_name() self.add_default_settings() def add_instance_name(self): """ Check if an InstanceName has been defined for this database. If not, create a random one! """ # See note above from .models import InvenTreeSetting """ Note: The "old" instance name was stored under the key 'InstanceName', but has now been renamed to 'INVENTREE_INSTANCE'. """ try: # Quick exit if a value already exists for 'inventree_instance' if InvenTreeSetting.objects.filter(key='INVENTREE_INSTANCE').exists(): return # Default instance name instance_name = InvenTreeSetting.get_default_value('INVENTREE_INSTANCE') # Use the old name if it exists if InvenTreeSetting.objects.filter(key='InstanceName').exists(): instance = InvenTreeSetting.objects.get(key='InstanceName') instance_name = instance.value # Delete the legacy key instance.delete() # Create new value InvenTreeSetting.objects.create( key='INVENTREE_INSTANCE', value=instance_name ) except (OperationalError, ProgrammingError, IntegrityError): # Migrations have not yet been applied - table does not exist pass def add_default_settings(self): """ Create all required settings, if they do not exist. """ from .models import InvenTreeSetting for key in InvenTreeSetting.GLOBAL_SETTINGS.keys(): try: settings = InvenTreeSetting.objects.filter(key__iexact=key) if settings.count() == 0: value = InvenTreeSetting.get_default_value(key) print(f"Creating default setting for {key} -> '{value}'") InvenTreeSetting.objects.create( key=key, value=value ) return elif settings.count() > 1: # Prevent multiple shadow copies of the same setting! for setting in settings[1:]: setting.delete() # Ensure that the key has the correct case setting = settings[0] if not setting.key == key: setting.key = key setting.save() except (OperationalError, ProgrammingError, IntegrityError): # Table might not yet exist pass
from django.apps import AppConfig from django.db.utils import OperationalError, ProgrammingError, IntegrityError class CommonConfig(AppConfig): name = 'common' def ready(self): pass
Remove code which automatically created settings objects on server launch
Remove code which automatically created settings objects on server launch
Python
mit
inventree/InvenTree,SchrodingersGat/InvenTree,inventree/InvenTree,inventree/InvenTree,inventree/InvenTree,SchrodingersGat/InvenTree,SchrodingersGat/InvenTree,SchrodingersGat/InvenTree
--- +++ @@ -6,87 +6,4 @@ name = 'common' def ready(self): - - """ Will be called when the Common app is first loaded """ - self.add_instance_name() - self.add_default_settings() - - def add_instance_name(self): - """ - Check if an InstanceName has been defined for this database. - If not, create a random one! - """ - - # See note above - from .models import InvenTreeSetting - - """ - Note: The "old" instance name was stored under the key 'InstanceName', - but has now been renamed to 'INVENTREE_INSTANCE'. - """ - - try: - - # Quick exit if a value already exists for 'inventree_instance' - if InvenTreeSetting.objects.filter(key='INVENTREE_INSTANCE').exists(): - return - - # Default instance name - instance_name = InvenTreeSetting.get_default_value('INVENTREE_INSTANCE') - - # Use the old name if it exists - if InvenTreeSetting.objects.filter(key='InstanceName').exists(): - instance = InvenTreeSetting.objects.get(key='InstanceName') - instance_name = instance.value - - # Delete the legacy key - instance.delete() - - # Create new value - InvenTreeSetting.objects.create( - key='INVENTREE_INSTANCE', - value=instance_name - ) - - except (OperationalError, ProgrammingError, IntegrityError): - # Migrations have not yet been applied - table does not exist - pass - - def add_default_settings(self): - """ - Create all required settings, if they do not exist. - """ - - from .models import InvenTreeSetting - - for key in InvenTreeSetting.GLOBAL_SETTINGS.keys(): - try: - settings = InvenTreeSetting.objects.filter(key__iexact=key) - - if settings.count() == 0: - value = InvenTreeSetting.get_default_value(key) - - print(f"Creating default setting for {key} -> '{value}'") - - InvenTreeSetting.objects.create( - key=key, - value=value - ) - - return - - elif settings.count() > 1: - # Prevent multiple shadow copies of the same setting! - for setting in settings[1:]: - setting.delete() - - # Ensure that the key has the correct case - setting = settings[0] - - if not setting.key == key: - setting.key = key - setting.save() - - except (OperationalError, ProgrammingError, IntegrityError): - # Table might not yet exist - pass + pass
f9dca979768ea17cee0993dac5bac4257bda623e
settings.py
settings.py
from settings_common import * DEBUG = TEMPLATE_DEBUG = True DATABASE_ENGINE = 'postgresql_psycopg2' DATABASE_NAME = 'daisyproducer_dev' DATABASE_USER = 'eglic' DATABASE_PASSWORD = '' DAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', 'tmp', 'pipeline-20090410') # debug toolbar INSTALLED_APPS += ('debug_toolbar',) MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',) INTERNAL_IPS = ('127.0.0.1',) DEBUG_TOOLBAR_CONFIG = {'INTERCEPT_REDIRECTS' : False}
from settings_common import * DEBUG = TEMPLATE_DEBUG = True DATABASE_ENGINE = 'postgresql_psycopg2' DATABASE_NAME = 'daisyproducer_dev' DATABASE_USER = 'eglic' DATABASE_PASSWORD = '' DAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', 'tmp', 'pipeline-20090410') # debug toolbar #INSTALLED_APPS += ('debug_toolbar',) #MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',) INTERNAL_IPS = ('127.0.0.1',) DEBUG_TOOLBAR_CONFIG = {'INTERCEPT_REDIRECTS' : False}
Comment out the debug tool bar
Comment out the debug tool bar
Python
agpl-3.0
sbsdev/daisyproducer,sbsdev/daisyproducer,sbsdev/daisyproducer,sbsdev/daisyproducer
--- +++ @@ -10,7 +10,7 @@ DAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', 'tmp', 'pipeline-20090410') # debug toolbar -INSTALLED_APPS += ('debug_toolbar',) -MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',) +#INSTALLED_APPS += ('debug_toolbar',) +#MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',) INTERNAL_IPS = ('127.0.0.1',) DEBUG_TOOLBAR_CONFIG = {'INTERCEPT_REDIRECTS' : False}
ec42a3cfcb491b265c87160ed9dae0005552acb4
tests/test_result.py
tests/test_result.py
from django.core import management import pytest from model_mommy import mommy import time from example.app.models import SimpleObject @pytest.mark.django_db def test_get(es_client): management.call_command("sync_es") test_object = mommy.make(SimpleObject) time.sleep(1) # Let the index refresh from_es = SimpleObject.search_objects.get(id=test_object.id) assert from_es.foo == test_object.foo assert from_es.bar == test_object.bar assert from_es.baz == test_object.baz assert from_es.__class__.__name__ == "SimpleObject_ElasticSearchResult" assert from_es.save is None
from django.core import management import pytest from model_mommy import mommy import time from example.app.models import SimpleObject, RelatableObject @pytest.mark.django_db def test_simple_get(es_client): management.call_command("sync_es") test_object = mommy.make(SimpleObject) time.sleep(1) # Let the index refresh from_es = SimpleObject.search_objects.get(id=test_object.id) assert from_es.foo == test_object.foo assert from_es.bar == test_object.bar assert from_es.baz == test_object.baz assert from_es.__class__.__name__ == "SimpleObject_ElasticSearchResult" assert from_es.save is None with pytest.raises(RelatableObject.DoesNotExist): RelatableObject.search_objects.get(id=test_object.id) @pytest.mark.django_db def test_related_get(es_client): management.call_command("sync_es") test_object = mommy.make(RelatableObject) time.sleep(1) # Let the index refresh from_es = RelatableObject.search_objects.get(id=test_object.id) assert from_es.foo == test_object.foo assert from_es.bar == test_object.bar assert from_es.baz == test_object.baz assert from_es.__class__.__name__ == "SimpleObject_ElasticSearchResult" assert from_es.save is None
Work on testing, bulk indexing, etc
Work on testing, bulk indexing, etc
Python
mit
theonion/djes
--- +++ @@ -3,11 +3,11 @@ from model_mommy import mommy import time -from example.app.models import SimpleObject +from example.app.models import SimpleObject, RelatableObject @pytest.mark.django_db -def test_get(es_client): +def test_simple_get(es_client): management.call_command("sync_es") @@ -20,3 +20,24 @@ assert from_es.baz == test_object.baz assert from_es.__class__.__name__ == "SimpleObject_ElasticSearchResult" assert from_es.save is None + + with pytest.raises(RelatableObject.DoesNotExist): + RelatableObject.search_objects.get(id=test_object.id) + + +@pytest.mark.django_db +def test_related_get(es_client): + + management.call_command("sync_es") + + test_object = mommy.make(RelatableObject) + time.sleep(1) # Let the index refresh + + from_es = RelatableObject.search_objects.get(id=test_object.id) + assert from_es.foo == test_object.foo + assert from_es.bar == test_object.bar + assert from_es.baz == test_object.baz + assert from_es.__class__.__name__ == "SimpleObject_ElasticSearchResult" + assert from_es.save is None + +
517bb590edb65baedc603d8ea64a5b6f5988f076
polyaxon/polyaxon/config_settings/scheduler/__init__.py
polyaxon/polyaxon/config_settings/scheduler/__init__.py
from polyaxon.config_settings.cors import * from polyaxon.config_settings.dirs import * from polyaxon.config_settings.k8s import * from polyaxon.config_settings.spawner import * from polyaxon.config_settings.registry import * from .apps import *
from polyaxon.config_settings.cors import * from polyaxon.config_settings.dirs import * from polyaxon.config_settings.k8s import * from polyaxon.config_settings.spawner import * from polyaxon.config_settings.registry import * from polyaxon.config_settings.volume_claims import * from .apps import *
Add volume claims to scheduler
Add volume claims to scheduler
Python
apache-2.0
polyaxon/polyaxon,polyaxon/polyaxon,polyaxon/polyaxon
--- +++ @@ -3,4 +3,5 @@ from polyaxon.config_settings.k8s import * from polyaxon.config_settings.spawner import * from polyaxon.config_settings.registry import * +from polyaxon.config_settings.volume_claims import * from .apps import *
39077720e2fcc340b0cc26a4720aa8d895c53263
src/txamqp/queue.py
src/txamqp/queue.py
# coding: utf-8 from twisted.internet.defer import DeferredQueue class Empty(Exception): pass class Closed(Exception): pass class TimeoutDeferredQueue(DeferredQueue): END = object() def _timeout(self, deferred): if not deferred.called: if deferred in self.waiting: self.waiting.remove(deferred) deferred.errback(Empty()) def _raiseIfClosed(self, result): if result == TimeoutDeferredQueue.END: self.put(TimeoutDeferredQueue.END) raise Closed() else: return result def get(self, timeout=None): deferred = DeferredQueue.get(self) deferred.addCallback(self._raiseIfClosed) if timeout: deferred.setTimeout(timeout, timeoutFunc=self._timeout) return deferred def close(self): self.put(TimeoutDeferredQueue.END)
# coding: utf-8 from twisted.internet.defer import DeferredQueue class Empty(Exception): pass class Closed(Exception): pass class TimeoutDeferredQueue(DeferredQueue): END = object() def _timeout(self, deferred): if not deferred.called: if deferred in self.waiting: self.waiting.remove(deferred) deferred.errback(Empty()) def _raiseIfClosed(self, result, call_id): if call_id is not None: call_id.cancel() if result == TimeoutDeferredQueue.END: self.put(TimeoutDeferredQueue.END) raise Closed() else: return result def get(self, timeout=None): deferred = DeferredQueue.get(self) call_id = None if timeout: from twisted.internet import reactor call_id = reactor.callLater(timeout, self._timeout, deferred) deferred.addCallback(self._raiseIfClosed, call_id) return deferred def close(self): self.put(TimeoutDeferredQueue.END)
Remove call to setTimeout and use callLater instead.
Remove call to setTimeout and use callLater instead.
Python
apache-2.0
williamsjj/txamqp,dotsent/txamqp,txamqp/txamqp
--- +++ @@ -17,7 +17,9 @@ self.waiting.remove(deferred) deferred.errback(Empty()) - def _raiseIfClosed(self, result): + def _raiseIfClosed(self, result, call_id): + if call_id is not None: + call_id.cancel() if result == TimeoutDeferredQueue.END: self.put(TimeoutDeferredQueue.END) @@ -28,10 +30,12 @@ def get(self, timeout=None): deferred = DeferredQueue.get(self) - deferred.addCallback(self._raiseIfClosed) + call_id = None + if timeout: + from twisted.internet import reactor + call_id = reactor.callLater(timeout, self._timeout, deferred) + deferred.addCallback(self._raiseIfClosed, call_id) - if timeout: - deferred.setTimeout(timeout, timeoutFunc=self._timeout) return deferred def close(self):
7eeb990644f387741ff4c217e1eaeddbe250988f
style_grader_main.py
style_grader_main.py
#!/usr/bin/python from style_grader_functions import * #TODO: Set up standard error to print properly def main(): student_file_names = get_arguments(sys.argv[1:]) sys.stderr = codecs.StreamReaderWriter(sys.stderr, codecs.getreader('utf8'), codecs.getwriter('utf8'), 'replace') rubric = StyleRubric() #rubric.reset_error_count() # Is this line necessary? operator_space_tracker = OperatorSpace() for filename in student_file_names: rubric.reset_for_new_file() # Fixes issue with multiple command-line arguments grade_student_file(filename, rubric, operator_space_tracker) #For debugging purposes only print "Total Errors: " + str(rubric.total_errors) for x, y in rubric.error_types.items(): print x, y #function called on each filename function(fileName, rubric) #print / send results if __name__ == '__main__': main()
#!/usr/bin/python from style_grader_functions import * #TODO: Set up standard error to print properly def main(): sys.stderr = codecs.StreamReaderWriter(sys.stderr, codecs.getreader('utf8'), codecs.getwriter('utf8'), 'replace') # Quick fix for now - ultimately this should be handled using argparse (TODO) if len(sys.argv) == 1: # No files were provided sys.stderr.write("Error: No files provided\n") # Should print usage info here, but argparse will autogen that - skipping this until that's decided sys.stderr.write("<Generic usage info>\n") student_file_names = get_arguments(sys.argv[1:]) rubric = StyleRubric() #rubric.reset_error_count() # Is this line necessary? operator_space_tracker = OperatorSpace() for filename in student_file_names: rubric.reset_for_new_file() # Fixes issue with multiple command-line arguments grade_student_file(filename, rubric, operator_space_tracker) #For debugging purposes only print "Total Errors: " + str(rubric.total_errors) for x, y in rubric.error_types.items(): print x, y #function called on each filename function(fileName, rubric) #print / send results if __name__ == '__main__': main()
Check at least one file was provided
Check at least one file was provided
Python
mit
vianuevm/cppStyle,vianuevm/cppStyle,vianuevm/cppStyle,vianuevm/cppStyle
--- +++ @@ -3,11 +3,19 @@ #TODO: Set up standard error to print properly def main(): - student_file_names = get_arguments(sys.argv[1:]) + sys.stderr = codecs.StreamReaderWriter(sys.stderr, codecs.getreader('utf8'), codecs.getwriter('utf8'), 'replace') + # Quick fix for now - ultimately this should be handled using argparse (TODO) + if len(sys.argv) == 1: + # No files were provided + sys.stderr.write("Error: No files provided\n") + # Should print usage info here, but argparse will autogen that - skipping this until that's decided + sys.stderr.write("<Generic usage info>\n") + + student_file_names = get_arguments(sys.argv[1:]) rubric = StyleRubric() #rubric.reset_error_count() # Is this line necessary?
baabb3a84418516e5a76a61b334b7879737b3d4b
goog/urls.py
goog/urls.py
from django.conf.urls.defaults import patterns, url urlpatterns = patterns( 'goog.views', url('^__goog__/(?P<path>.*)$', 'serve_closure', name='goog_serve_closure'), # FIXME(andi): That's a bit ugly to cover third_party as an URL... url('^third_party/(?P<path>.*)$', 'serve_closure_thirdparty', name='goog_serve_closure_tp'), )
try: from django.conf.urls.defaults import patterns, url except ImportError: # Django >= 1.6 from django.conf.urls import patterns, url urlpatterns = patterns( 'goog.views', url('^__goog__/(?P<path>.*)$', 'serve_closure', name='goog_serve_closure'), # FIXME(andi): That's a bit ugly to cover third_party as an URL... url('^third_party/(?P<path>.*)$', 'serve_closure_thirdparty', name='goog_serve_closure_tp'), )
Fix imports for Django >= 1.6
Fix imports for Django >= 1.6
Python
bsd-3-clause
andialbrecht/django-goog
--- +++ @@ -1,4 +1,7 @@ -from django.conf.urls.defaults import patterns, url +try: + from django.conf.urls.defaults import patterns, url +except ImportError: # Django >= 1.6 + from django.conf.urls import patterns, url urlpatterns = patterns( 'goog.views',
ed46ee16ed1b8efcee3697d3da909f72b0755a13
webcomix/tests/test_docker.py
webcomix/tests/test_docker.py
import docker from webcomix.docker import DockerManager def test_no_javascript_spawns_no_container(): manager = DockerManager(False) manager.__enter__() manager.client = docker.from_env() assert manager._get_container() is None def test_javascript_spawns_container(): manager = DockerManager(True) manager.__enter__() assert manager._get_container() is not None manager.__exit__(None, None, None) def test_javascript_exit_removes_container(): manager = DockerManager(True) manager.__enter__() manager.__exit__(None, None, None) assert manager._get_container() is None
import docker import pytest from webcomix.docker import DockerManager, CONTAINER_NAME @pytest.fixture def cleanup_container(test): yield None client = docker.from_env() for container in client.containers().list(): if container.attrs["Config"]["Image"] == CONTAINER_NAME: container.kill() def test_no_javascript_spawns_no_container(cleanup_container): manager = DockerManager(False) manager.__enter__() manager.client = docker.from_env() assert manager._get_container() is None def test_javascript_spawns_container(cleanup_container): manager = DockerManager(True) manager.__enter__() assert manager._get_container() is not None manager.__exit__(None, None, None) def test_javascript_exit_removes_container(cleanup_container): manager = DockerManager(True) manager.__enter__() manager.__exit__(None, None, None) assert manager._get_container() is None
Add test fixture for docker tests
Add test fixture for docker tests
Python
mit
J-CPelletier/webcomix,J-CPelletier/webcomix
--- +++ @@ -1,20 +1,29 @@ import docker +import pytest -from webcomix.docker import DockerManager +from webcomix.docker import DockerManager, CONTAINER_NAME -def test_no_javascript_spawns_no_container(): +@pytest.fixture +def cleanup_container(test): + yield None + client = docker.from_env() + for container in client.containers().list(): + if container.attrs["Config"]["Image"] == CONTAINER_NAME: + container.kill() + +def test_no_javascript_spawns_no_container(cleanup_container): manager = DockerManager(False) manager.__enter__() manager.client = docker.from_env() assert manager._get_container() is None -def test_javascript_spawns_container(): +def test_javascript_spawns_container(cleanup_container): manager = DockerManager(True) manager.__enter__() assert manager._get_container() is not None manager.__exit__(None, None, None) -def test_javascript_exit_removes_container(): +def test_javascript_exit_removes_container(cleanup_container): manager = DockerManager(True) manager.__enter__() manager.__exit__(None, None, None)
16b07dd961cbe55ee452ed6057048ec452ffbd72
custom/icds/management/commands/copy_icds_app.py
custom/icds/management/commands/copy_icds_app.py
from __future__ import absolute_import, print_function, unicode_literals from django.core.management import BaseCommand from corehq.apps.app_manager.dbaccessors import get_build_doc_by_version, wrap_app from corehq.apps.app_manager.models import import_app class Command(BaseCommand): help = "Make a copy of a specific version of an application on the same domain" def add_arguments(self, parser): parser.add_argument('domain') parser.add_argument('app_id') parser.add_argument('version') parser.add_argument('new_name') def handle(self, domain, app_id, version, new_name, **options): old_app = get_build_doc_by_version(domain, app_id, version) if not old_app: raise Exception("No app found with id '{}' and version '{}', on '{}'" .format(app_id, version, domain)) old_app = wrap_app(old_app) old_app.convert_build_to_app() new_app = import_app(old_app.to_json(), domain, source_properties={'name': new_name})
from __future__ import absolute_import, print_function, unicode_literals from django.core.management import BaseCommand from corehq.apps.app_manager.dbaccessors import get_build_doc_by_version, wrap_app from corehq.apps.app_manager.models import import_app class Command(BaseCommand): help = "Make a copy of a specific version of an application on the same domain" def add_arguments(self, parser): parser.add_argument('domain') parser.add_argument('app_id') parser.add_argument('version') parser.add_argument('new_name') def handle(self, domain, app_id, version, new_name, **options): old_app = get_build_doc_by_version(domain, app_id, version) if not old_app: raise Exception("No app found with id '{}' and version '{}', on '{}'" .format(app_id, version, domain)) old_app = wrap_app(old_app) old_app.convert_build_to_app() new_app = import_app(old_app.to_json(), domain, source_properties={'name': new_name}) old_to_new = get_old_to_new_config_ids(old_app, new_app) for form in new_app.get_forms(): for old_id, new_id in old_to_new: form.source = form.source.replace(old_id, new_id) new_app.save() def get_old_to_new_config_ids(old_app, new_app): return [ (old_config.uuid, new_config.uuid) for old_module, new_module in zip(old_app.get_report_modules(), new_app.get_report_modules()) for old_config, new_config in zip(old_module.report_configs, new_module.report_configs) ]
Replace old config IDs with the new ones
Replace old config IDs with the new ones
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
--- +++ @@ -23,3 +23,18 @@ old_app = wrap_app(old_app) old_app.convert_build_to_app() new_app = import_app(old_app.to_json(), domain, source_properties={'name': new_name}) + + old_to_new = get_old_to_new_config_ids(old_app, new_app) + for form in new_app.get_forms(): + for old_id, new_id in old_to_new: + form.source = form.source.replace(old_id, new_id) + + new_app.save() + + +def get_old_to_new_config_ids(old_app, new_app): + return [ + (old_config.uuid, new_config.uuid) + for old_module, new_module in zip(old_app.get_report_modules(), new_app.get_report_modules()) + for old_config, new_config in zip(old_module.report_configs, new_module.report_configs) + ]
6f356a94c56053b47fb38670a93e04f46740f21e
tartpy/eventloop.py
tartpy/eventloop.py
""" Very basic implementation of an event loop ========================================== The eventloop is a singleton to schedule and run events. Exports ------- - ``EventLoop``: the basic eventloop """ import asyncio import queue import sched import threading import time from .singleton import Singleton class EventLoop(object, metaclass=Singleton): def __init__(self): self.loop = asyncio.get_event_loop() self.do = self.sync_do def sync_do(self, f, *args, **kwargs): f(*args, **kwargs) def thread_do(self, f, *args, **kwargs): self.loop.call_soon_threadsafe(f, *args, **kwargs) def schedule(self, target, event): self.do(self.loop.call_soon, event) def later(self, delay, event): self.do(self.loop.call_later, delay, event) def run(self): self.do = self.sync_do self.loop.run_forever() def run_once(self): self.stop_later() self.run() def run_in_thread(self): self.do = self.thread_do self.thread = threading.Thread(target=self.loop.run_forever, name='asyncio_event_loop') self.thread.daemon = True self.thread.start() def stop(self): self.loop.stop() def stop_later(self): self.do = self.sync_do self.schedule(self, self.stop)
""" Very basic implementation of an event loop ========================================== The eventloop is a singleton to schedule and run events. Exports ------- - ``EventLoop``: the basic eventloop """ import asyncio import queue import sched import threading import time from .singleton import Singleton class EventLoop(object, metaclass=Singleton): def __init__(self): self.loop = asyncio.get_event_loop() self.do = self.sync_do def sync_do(self, f, *args, **kwargs): f(*args, **kwargs) def thread_do(self, f, *args, **kwargs): self.loop.call_soon_threadsafe(f, *args, **kwargs) def schedule(self, target, event): self.do(self.loop.call_soon, event) def later(self, delay, event): self.do(self.loop.call_later, delay, event) def run(self): self.do = self.sync_do self.loop.run_forever() def run_once(self): self.stop_later() self.run() def run_in_thread(self): self.do = self.thread_do self.thread = threading.Thread(target=self.loop.run_forever, name='asyncio_event_loop') self.thread.daemon = True self.thread.start() def stop(self): self.thread_do(self.loop.stop) def stop_later(self): self.do = self.sync_do self.schedule(self, self.stop)
Make sure that 'stop' works from everywhere
Make sure that 'stop' works from everywhere
Python
mit
waltermoreira/tartpy
--- +++ @@ -55,7 +55,7 @@ self.thread.start() def stop(self): - self.loop.stop() + self.thread_do(self.loop.stop) def stop_later(self): self.do = self.sync_do
39b5f794503149351d03879083d336dfe5f2351b
openprescribing/frontend/tests/test_api_utils.py
openprescribing/frontend/tests/test_api_utils.py
from django.test import TestCase from django.db import OperationalError class ApiTestUtils(TestCase): def test_db_timeout(self): from api.view_utils import db_timeout @db_timeout(1) def do_long_running_query(): from django.db import connection cursor = connection.cursor() cursor.execute("select pg_sleep(0.01);") self.assertRaises(OperationalError, do_long_running_query)
from django.test import TestCase from django.db import OperationalError class ApiTestUtils(TestCase): def test_db_timeout(self): from api.view_utils import db_timeout @db_timeout(1) def do_long_running_query(): from django.db import connection cursor = connection.cursor() cursor.execute("select pg_sleep(0.01);") self.assertRaises(OperationalError, do_long_running_query) def test_param_to_list(self): from api.view_utils import param_to_list self.assertEquals(param_to_list('foo'), ['foo']) self.assertEquals(param_to_list('foo,bar'), ['foo', 'bar']) self.assertEquals(param_to_list('foo,bar'), ['foo', 'bar']) self.assertEquals(param_to_list(None), []) self.assertEquals(param_to_list([]), [])
Add a missing test for param-parsing
Add a missing test for param-parsing
Python
mit
ebmdatalab/openprescribing,ebmdatalab/openprescribing,ebmdatalab/openprescribing,annapowellsmith/openpresc,annapowellsmith/openpresc,annapowellsmith/openpresc,ebmdatalab/openprescribing,annapowellsmith/openpresc
--- +++ @@ -12,3 +12,12 @@ cursor = connection.cursor() cursor.execute("select pg_sleep(0.01);") self.assertRaises(OperationalError, do_long_running_query) + + def test_param_to_list(self): + from api.view_utils import param_to_list + + self.assertEquals(param_to_list('foo'), ['foo']) + self.assertEquals(param_to_list('foo,bar'), ['foo', 'bar']) + self.assertEquals(param_to_list('foo,bar'), ['foo', 'bar']) + self.assertEquals(param_to_list(None), []) + self.assertEquals(param_to_list([]), [])
bed98e0769d067857dadf69079dc049d76d65fd0
kitchen/lib/__init__.py
kitchen/lib/__init__.py
import os import json from kitchen.settings import KITCHEN_LOCATION def load_nodes(): retval = {} nodes_dir = os.path.join(KITCHEN_LOCATION, 'nodes') for filename in os.listdir(nodes_dir): f = open(os.path.join(nodes_dir, filename), 'r') retval[filename[:-5]] = json.load(f) f.close() return retval
Add function to load data from nodes files
Add function to load data from nodes files
Python
apache-2.0
edelight/kitchen,edelight/kitchen,edelight/kitchen,edelight/kitchen
--- +++ @@ -0,0 +1,13 @@ +import os +import json + +from kitchen.settings import KITCHEN_LOCATION + +def load_nodes(): + retval = {} + nodes_dir = os.path.join(KITCHEN_LOCATION, 'nodes') + for filename in os.listdir(nodes_dir): + f = open(os.path.join(nodes_dir, filename), 'r') + retval[filename[:-5]] = json.load(f) + f.close() + return retval
27467c09abe01a6e6a2b66f9f7553bb36cb8a977
uploader/uploader.py
uploader/uploader.py
#!/usr/bin/python3 from __future__ import print_function import os import time import subprocess import sys WAIT = 30 def main(): directory = sys.argv[1] url = os.environ['RSYNC_URL'] while True: fnames = list(f for f in os.listdir(directory) if f.endswith('.warc.gz')) if len(fnames): fname = os.path.join(directory, fnames[0]) print("Uploading %r" % (fname,)) exit = subprocess.call(["ionice", "-c", "2", "-n", "0", "rsync", "-av", "--timeout=300", "--contimeout=300", "--progress", fname, url]) if exit == 0: print("Removing %r" % (fname,)) os.remove(fname) else: print("Nothing to upload") print("Waiting %d seconds" % (WAIT,)) time.sleep(WAIT) if __name__ == '__main__': main()
#!/usr/bin/python3 from __future__ import print_function import os import time import subprocess import sys WAIT = 30 def main(): directory = sys.argv[1] url = os.environ['RSYNC_URL'] while True: fnames = sorted(list(f for f in os.listdir(directory) if f.endswith('.warc.gz'))) if len(fnames): fname = os.path.join(directory, fnames[0]) print("Uploading %r" % (fname,)) exit = subprocess.call(["ionice", "-c", "2", "-n", "0", "rsync", "-av", "--timeout=300", "--contimeout=300", "--progress", fname, url]) if exit == 0: print("Removing %r" % (fname,)) os.remove(fname) else: print("Nothing to upload") print("Waiting %d seconds" % (WAIT,)) time.sleep(WAIT) if __name__ == '__main__': main()
Sort files before choosing one to upload
Sort files before choosing one to upload
Python
mit
Frogging101/ArchiveBot,emijrp/ArchiveBot,emijrp/ArchiveBot,Frogging101/ArchiveBot,Asparagirl/ArchiveBot,ArchiveTeam/ArchiveBot,JesseWeinstein/ArchiveBot,falconkirtaran/ArchiveBot,emijrp/ArchiveBot,Frogging101/ArchiveBot,JesseWeinstein/ArchiveBot,Asparagirl/ArchiveBot,Frogging101/ArchiveBot,emijrp/ArchiveBot,emijrp/ArchiveBot,JesseWeinstein/ArchiveBot,JesseWeinstein/ArchiveBot,ArchiveTeam/ArchiveBot,falconkirtaran/ArchiveBot,falconkirtaran/ArchiveBot,falconkirtaran/ArchiveBot,Frogging101/ArchiveBot,Asparagirl/ArchiveBot,ArchiveTeam/ArchiveBot,ArchiveTeam/ArchiveBot,Asparagirl/ArchiveBot,Asparagirl/ArchiveBot,falconkirtaran/ArchiveBot,JesseWeinstein/ArchiveBot,ArchiveTeam/ArchiveBot
--- +++ @@ -13,7 +13,7 @@ directory = sys.argv[1] url = os.environ['RSYNC_URL'] while True: - fnames = list(f for f in os.listdir(directory) if f.endswith('.warc.gz')) + fnames = sorted(list(f for f in os.listdir(directory) if f.endswith('.warc.gz'))) if len(fnames): fname = os.path.join(directory, fnames[0]) print("Uploading %r" % (fname,))
d108f090f198cba47083225c0de46e77b22ab5cc
serfclient/__init__.py
serfclient/__init__.py
from pkg_resources import get_distribution __version__ = get_distribution('serfclient').version from serfclient.client import SerfClient
from pkg_resources import get_distribution from serfclient.client import SerfClient __version__ = get_distribution('serfclient').version
Move module level import to top of file (PEP8)
Move module level import to top of file (PEP8) Error: E402 module level import not at top of file
Python
mit
charleswhchan/serfclient-py,KushalP/serfclient-py
--- +++ @@ -1,5 +1,4 @@ from pkg_resources import get_distribution +from serfclient.client import SerfClient __version__ = get_distribution('serfclient').version - -from serfclient.client import SerfClient
b26aaf9bdc80760236d4369f67ea803becc733b7
test_squarespace.py
test_squarespace.py
# coding=UTF-8 from squarespace import Squarespace def test_squarespace(): store = Squarespace('test') assert store.api_key == 'test' assert store.useragent == 'Squarespace python API v0.0.1 by Zach White.' def test_squarespace_useragent(): store = Squarespace('test') store.useragent = 'Hello, World!' assert store.useragent == 'Hello, World!' assert store._useragent == 'Hello, World!' assert store.http.headers['User-Agent'] == 'Hello, World!'
# coding=UTF-8 from squarespace import Squarespace def test_squarespace(): store = Squarespace('test') assert store.api_key == 'test' assert store.useragent == 'Squarespace python API v0.0.2 by Zach White.' def test_squarespace_useragent(): store = Squarespace('test') store.useragent = 'Hello, World!' assert store.useragent == 'Hello, World!' assert store._useragent == 'Hello, World!' assert store.http.headers['User-Agent'] == 'Hello, World!'
Increment the version in the test too
Increment the version in the test too
Python
mit
skullydazed/squarespace-python,skullydazed/squarespace-python
--- +++ @@ -5,7 +5,7 @@ def test_squarespace(): store = Squarespace('test') assert store.api_key == 'test' - assert store.useragent == 'Squarespace python API v0.0.1 by Zach White.' + assert store.useragent == 'Squarespace python API v0.0.2 by Zach White.' def test_squarespace_useragent():
375fd952e4495a07cb2031c7d380bdb4a535defc
tests/test_dimension.py
tests/test_dimension.py
from devito import SubsampledDimension, Grid, TimeFunction, Eq, Operator from devito.tools import pprint def test_subsampled_dimension(): nt = 10 grid = Grid(shape=(11, 11)) x, y = grid.dimensions time = grid.time_dim t = grid.stepping_dim time_subsampled = SubsampledDimension('t_sub', parent=time, factor=4) u = TimeFunction(name='u', grid=grid) u2 = TimeFunction(name='u2', grid=grid, save=nt) assert(t in u.indices) u_s = TimeFunction(name='u_s', grid=grid, time_dim=time_subsampled) assert(time_subsampled in u_s.indices) fwd_eqn = Eq(u.indexed[t+1, x, y], u.indexed[t, x, y] + 1.) fwd_eqn_2 = Eq(u2.indexed[time+1, x, y], u2.indexed[time, x, y] + 1.) save_eqn = Eq(u_s, u) #fwd_op = Operator([fwd_eqn]) fwd_op = Operator([fwd_eqn, fwd_eqn_2, save_eqn]) pprint(fwd_op) print(fwd_op)
from devito import SubsampledDimension, Grid, TimeFunction, Eq, Operator from devito.tools import pprint def test_subsampled_dimension(): nt = 10 grid = Grid(shape=(11, 11)) x, y = grid.dimensions time = grid.time_dim t = grid.stepping_dim time_subsampled = SubsampledDimension('t_sub', parent=time, factor=4) u = TimeFunction(name='u', grid=grid) u2 = TimeFunction(name='u2', grid=grid, save=nt) assert(t in u.indices) u_s = TimeFunction(name='u_s', grid=grid, time_dim=time_subsampled) assert(time_subsampled in u_s.indices) fwd_eqn = Eq(u.indexed[t+1, x, y], u.indexed[t, x, y] + 1.) fwd_eqn_2 = Eq(u2.indexed[time+1, x, y], u2.indexed[time, x, y] + 1.) save_eqn = Eq(u_s, u) #fwd_op = Operator([fwd_eqn]) fwd_op = Operator([fwd_eqn, fwd_eqn_2, save_eqn]) #pprint(fwd_op) #print(fwd_op)
Change from is_Stepping to is_Derived wherever appropriate
Change from is_Stepping to is_Derived wherever appropriate
Python
mit
opesci/devito,opesci/devito
--- +++ @@ -19,5 +19,5 @@ save_eqn = Eq(u_s, u) #fwd_op = Operator([fwd_eqn]) fwd_op = Operator([fwd_eqn, fwd_eqn_2, save_eqn]) - pprint(fwd_op) - print(fwd_op) + #pprint(fwd_op) + #print(fwd_op)
01a8fcb70ea75d854aaf16547b837d861750c160
tilequeue/queue/file.py
tilequeue/queue/file.py
from tilequeue.tile import serialize_coord, deserialize_coord, CoordMessage import threading class OutputFileQueue(object): def __init__(self, fp): self.fp = fp self.lock = threading.RLock() def enqueue(self, coord): with self.lock: payload = serialize_coord(coord) self.fp.write(payload + '\n') def enqueue_batch(self, coords): n = 0 for coord in coords: self.enqueue(coord) n += 1 return n, 0 def read(self, max_to_read=1, timeout_seconds=20): with self.lock: coords = [] for _ in range(max_to_read): try: coord = next(self.fp) except StopIteration: break coords.append(CoordMessage(deserialize_coord(coord), None)) return coords def job_done(self, coord_message): pass def clear(self): with self.lock: self.fp.seek(0) self.fp.truncate() return -1 def close(self): with self.lock: remaining_queue = "".join([ln for ln in self.fp]) self.clear() self.fp.write(remaining_queue) self.fp.close()
from tilequeue.tile import serialize_coord, deserialize_coord, CoordMessage import threading class OutputFileQueue(object): def __init__(self, fp): self.fp = fp self.lock = threading.RLock() def enqueue(self, coord): with self.lock: payload = serialize_coord(coord) self.fp.write(payload + '\n') def enqueue_batch(self, coords): n = 0 for coord in coords: self.enqueue(coord) n += 1 return n, 0 def read(self, max_to_read=1, timeout_seconds=20): with self.lock: coords = [] for _ in range(max_to_read): coord = self.fp.readline() if coord: coords.append(CoordMessage(deserialize_coord(coord), None)) else: break return coords def job_done(self, coord_message): pass def clear(self): with self.lock: self.fp.seek(0) self.fp.truncate() return -1 def close(self): with self.lock: remaining_queue = ''.join([ln for ln in self.fp]) self.clear() self.fp.write(remaining_queue) self.fp.close()
Use readline() instead of next() to detect changes.
Use readline() instead of next() to detect changes. tilequeue/queue/file.py -`readline()` will pick up new lines appended to the file, whereas `next()` will not since the iterator will just hit `StopIteration` and stop generating new lines. Use `readline()` instead, then, since it might be desirable to append something to the queue file and have tilequeue detect the new input without having to restart.
Python
mit
tilezen/tilequeue,mapzen/tilequeue
--- +++ @@ -24,11 +24,11 @@ with self.lock: coords = [] for _ in range(max_to_read): - try: - coord = next(self.fp) - except StopIteration: + coord = self.fp.readline() + if coord: + coords.append(CoordMessage(deserialize_coord(coord), None)) + else: break - coords.append(CoordMessage(deserialize_coord(coord), None)) return coords @@ -43,7 +43,7 @@ def close(self): with self.lock: - remaining_queue = "".join([ln for ln in self.fp]) + remaining_queue = ''.join([ln for ln in self.fp]) self.clear() self.fp.write(remaining_queue) self.fp.close()
5456ae0af9ad83b8e0339c671ce8954bb48d62cf
database.py
database.py
from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker from sqlalchemy.ext.declarative import declarative_base import config as cfg class DB(object): engine = None db_session = None Base = declarative_base() def __init__(self, dbstring): self.engine = create_engine(dbstring, convert_unicode=True) self.db_session = scoped_session(sessionmaker(autocommit=False, autoflush=False, bind=self.engine)) self.Base.query = self.db_session.query_property() def get_session(self): return self.db_session def get_base(self): return self.Base
from sqlalchemy import create_engine from sqlalchemy.orm import scoped_session, sessionmaker, class_mapper from sqlalchemy.ext.declarative import declarative_base import config as cfg class DB(object): engine = None db_session = None Base = declarative_base() def __init__(self, dbstring): self.engine = create_engine(dbstring, convert_unicode=True) self.db_session = scoped_session(sessionmaker(autocommit=False, autoflush=False, bind=self.engine)) self.Base.query = self.db_session.query_property() def get_session(self): return self.db_session def get_base(self): return self.Base class ImposterBase(object): """ Mixin class to provide additional generic functions for the sqlalchemy models """ def to_dict(obj): """Return dict containing all object data""" return dict((col.name, unicode(getattr(obj, col.name))) for col in class_mapper(obj.__class__).mapped_table.c) def get_public_dict(obj): """Return dict containing only public object data""" return dict((col.name, unicode(getattr(obj, col.name))) for col in obj.__class__.__public_columns__)
Add ImposterBase mixin class so we can add methods and properties to the sqlalchemy based models
Add ImposterBase mixin class so we can add methods and properties to the sqlalchemy based models
Python
bsd-2-clause
jkossen/imposter,jkossen/imposter
--- +++ @@ -1,5 +1,5 @@ from sqlalchemy import create_engine -from sqlalchemy.orm import scoped_session, sessionmaker +from sqlalchemy.orm import scoped_session, sessionmaker, class_mapper from sqlalchemy.ext.declarative import declarative_base import config as cfg @@ -20,3 +20,16 @@ def get_base(self): return self.Base + +class ImposterBase(object): + """ Mixin class to provide additional generic functions for the sqlalchemy models """ + + def to_dict(obj): + """Return dict containing all object data""" + return dict((col.name, unicode(getattr(obj, col.name))) + for col in class_mapper(obj.__class__).mapped_table.c) + + def get_public_dict(obj): + """Return dict containing only public object data""" + return dict((col.name, unicode(getattr(obj, col.name))) + for col in obj.__class__.__public_columns__)
be1e31c78f17961851d41dea11cd912d237cf5fb
lib/rapidsms/message.py
lib/rapidsms/message.py
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 import copy class Message(object): def __init__(self, backend, caller=None, text=None): self._backend = backend self.caller = caller self.text = text # initialize some empty attributes self.received = None self.sent = None self.responses = [] def __unicode__(self): return self.text @property def backend(self): # backend is read-only, since it's an # immutable property of this object return self._backend def send(self): """Send this message via self.backend, returning True if the message was sent successfully.""" return self.backend.router.outgoing(self) def flush_responses (self): for response in self.responses: response.send() def respond(self, text): """Send the given text back to the original caller of this message on the same route that it came in on""" if self.caller: response = copy.copy(self) response.text = text self.responses.append(response) return True else: return False
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 import copy class Message(object): def __init__(self, backend, caller=None, text=None): self._backend = backend self.caller = caller self.text = text self.responses = [] def __unicode__(self): return self.text @property def backend(self): # backend is read-only, since it's an # immutable property of this object return self._backend def send(self): """Send this message via self.backend, returning True if the message was sent successfully.""" return self.backend.router.outgoing(self) def flush_responses (self): for response in self.responses: response.send() self.responses.remove(response) def respond(self, text): """Send the given text back to the original caller of this message on the same route that it came in on""" if self.caller: response = copy.copy(self) response.text = text self.responses.append(response) return True else: return False
Remove unused attributes; also, empty responses after it's flushed.
Remove unused attributes; also, empty responses after it's flushed.
Python
bsd-3-clause
unicefuganda/edtrac,caktus/rapidsms,eHealthAfrica/rapidsms,dimagi/rapidsms,catalpainternational/rapidsms,peterayeni/rapidsms,peterayeni/rapidsms,ken-muturi/rapidsms,ehealthafrica-ci/rapidsms,rapidsms/rapidsms-core-dev,eHealthAfrica/rapidsms,lsgunth/rapidsms,dimagi/rapidsms-core-dev,ehealthafrica-ci/rapidsms,catalpainternational/rapidsms,peterayeni/rapidsms,eHealthAfrica/rapidsms,catalpainternational/rapidsms,rapidsms/rapidsms-core-dev,unicefuganda/edtrac,lsgunth/rapidsms,dimagi/rapidsms-core-dev,peterayeni/rapidsms,lsgunth/rapidsms,ken-muturi/rapidsms,ken-muturi/rapidsms,dimagi/rapidsms,catalpainternational/rapidsms,unicefuganda/edtrac,caktus/rapidsms,caktus/rapidsms,ehealthafrica-ci/rapidsms,lsgunth/rapidsms
--- +++ @@ -8,10 +8,6 @@ self._backend = backend self.caller = caller self.text = text - - # initialize some empty attributes - self.received = None - self.sent = None self.responses = [] def __unicode__(self): @@ -31,6 +27,7 @@ def flush_responses (self): for response in self.responses: response.send() + self.responses.remove(response) def respond(self, text): """Send the given text back to the original caller of this
536575db87968014f75d2ad68456c3684d6c92de
auditlog/__manifest__.py
auditlog/__manifest__.py
# -*- coding: utf-8 -*- # © 2015 ABF OSIELL <http://osiell.com> # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). { 'name': "Audit Log", 'version': "9.0.1.0.0", 'author': "ABF OSIELL,Odoo Community Association (OCA)", 'license': "AGPL-3", 'website': "http://www.osiell.com", 'category': "Tools", 'depends': [ 'base', ], 'data': [ 'security/ir.model.access.csv', 'data/ir_cron.xml', 'views/auditlog_view.xml', 'views/http_session_view.xml', 'views/http_request_view.xml', ], 'images': [], 'application': True, 'installable': True, 'pre_init_hook': 'pre_init_hook', }
# -*- coding: utf-8 -*- # © 2015 ABF OSIELL <http://osiell.com> # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). { 'name': "Audit Log", 'version': "9.0.1.0.0", 'author': "ABF OSIELL,Odoo Community Association (OCA)", 'license': "AGPL-3", 'website': "http://www.osiell.com", 'category': "Tools", 'depends': [ 'base', ], 'data': [ 'security/ir.model.access.csv', 'data/ir_cron.xml', 'views/auditlog_view.xml', 'views/http_session_view.xml', 'views/http_request_view.xml', ], 'images': [], 'application': True, 'installable': True, }
Remove pre_init_hook reference from openerp, no pre_init hook exists any more
auditlog: Remove pre_init_hook reference from openerp, no pre_init hook exists any more
Python
agpl-3.0
brain-tec/server-tools,brain-tec/server-tools,bmya/server-tools,brain-tec/server-tools,bmya/server-tools,bmya/server-tools
--- +++ @@ -22,5 +22,4 @@ 'images': [], 'application': True, 'installable': True, - 'pre_init_hook': 'pre_init_hook', }
7f2e91064eabc020cbe660639713278fc187a034
tests/test_result.py
tests/test_result.py
import pytest from serfclient import result class TestSerfResult(object): def test_initialises_to_none(self): r = result.SerfResult() assert r.head is None assert r.body is None def test_provides_a_pretty_printed_form_for_repl_use(self): r = result.SerfResult(head={"a": 1}, body=('foo', 'bar')) assert str(r) == \ "SerfResult<head={'a': 1},body=('foo', 'bar')>" def test_can_convert_to_list(self): r = result.SerfResult(head=1, body=2) assert sorted(list(r)) == [1, 2] def test_can_convert_to_tuple(self): r = result.SerfResult(head=1, body=2) assert sorted(tuple(r)) == [1, 2]
from serfclient import result class TestSerfResult(object): def test_initialises_to_none(self): r = result.SerfResult() assert r.head is None assert r.body is None def test_provides_a_pretty_printed_form_for_repl_use(self): r = result.SerfResult(head={"a": 1}, body=('foo', 'bar')) assert str(r) == \ "SerfResult<head={'a': 1},body=('foo', 'bar')>" def test_can_convert_to_list(self): r = result.SerfResult(head=1, body=2) assert sorted(list(r)) == [1, 2] def test_can_convert_to_tuple(self): r = result.SerfResult(head=1, body=2) assert sorted(tuple(r)) == [1, 2]
Remove unused import of pytest
Remove unused import of pytest
Python
mit
charleswhchan/serfclient-py,KushalP/serfclient-py
--- +++ @@ -1,5 +1,3 @@ -import pytest - from serfclient import result
a9cd0a385253cef42d03d6a45e81ef4dd582e9de
base/settings/testing.py
base/settings/testing.py
# -*- coding: utf-8 -*- from .base import Base as Settings class Testing(Settings): # Database Configuration. # -------------------------------------------------------------------------- DATABASES = { 'default': { 'ENGINE': 'django.db.backends.postgresql_psycopg2', 'NAME': 'test', } } # django-celery. # -------------------------------------------------------------------------- Settings.INSTALLED_APPS += ['kombu.transport.django', 'djcelery'] BROKER_URL = 'django://' # django-haystack. # -------------------------------------------------------------------------- HAYSTACK_CONNECTIONS = { 'default': { 'ENGINE': 'haystack.backends.simple_backend.SimpleEngine', }, } # Media Storage Configuration. # -------------------------------------------------------------------------- # Amazon Web Services AWS_STORAGE_BUCKET_NAME = 'test-bucket' # django-s3-folder-storage DEFAULT_S3_PATH = 'media' STATIC_S3_PATH = 'static' CDN_DOMAIN = 'cdn.example.net' MEDIA_URL = 'https://%s/%s/' % (CDN_DOMAIN, DEFAULT_S3_PATH) STATIC_URL = 'https://%s/%s/' % (CDN_DOMAIN, STATIC_S3_PATH) # Authentication Configuration. # -------------------------------------------------------------------------- HELLO_BASE_CLIENT_ID = 'client-id' HELLO_BASE_CLIENT_SECRET = 'client-secret' OAUTH_AUTHORIZATION_URL = 'https://testserver/oauth/authorize/' OAUTH_TOKEN_URL = 'https://testserver/oauth/token/'
# -*- coding: utf-8 -*- from .base import Base as Settings class Testing(Settings): # Database Configuration. # -------------------------------------------------------------------------- DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': ':memory:', } } # django-celery. # -------------------------------------------------------------------------- Settings.INSTALLED_APPS += ['kombu.transport.django', 'djcelery'] BROKER_URL = 'django://' # django-haystack. # -------------------------------------------------------------------------- HAYSTACK_CONNECTIONS = { 'default': { 'ENGINE': 'haystack.backends.simple_backend.SimpleEngine', }, } # Media Storage Configuration. # -------------------------------------------------------------------------- # Amazon Web Services AWS_STORAGE_BUCKET_NAME = 'test-bucket' # django-s3-folder-storage DEFAULT_S3_PATH = 'media' STATIC_S3_PATH = 'static' CDN_DOMAIN = 'cdn.example.net' MEDIA_URL = 'https://%s/%s/' % (CDN_DOMAIN, DEFAULT_S3_PATH) STATIC_URL = 'https://%s/%s/' % (CDN_DOMAIN, STATIC_S3_PATH) # Authentication Configuration. # -------------------------------------------------------------------------- HELLO_BASE_CLIENT_ID = 'client-id' HELLO_BASE_CLIENT_SECRET = 'client-secret' OAUTH_AUTHORIZATION_URL = 'https://testserver/oauth/authorize/' OAUTH_TOKEN_URL = 'https://testserver/oauth/token/'
Use SQLite in an attempt to speed up the tests.
Use SQLite in an attempt to speed up the tests.
Python
apache-2.0
hello-base/web,hello-base/web,hello-base/web,hello-base/web
--- +++ @@ -7,8 +7,8 @@ # -------------------------------------------------------------------------- DATABASES = { 'default': { - 'ENGINE': 'django.db.backends.postgresql_psycopg2', - 'NAME': 'test', + 'ENGINE': 'django.db.backends.sqlite3', + 'NAME': ':memory:', } }
211ee03b811cb196dd9f36026fcfc6e75dda2ec6
byceps/config_defaults.py
byceps/config_defaults.py
""" byceps.config_defaults ~~~~~~~~~~~~~~~~~~~~~~ Default configuration values :Copyright: 2006-2021 Jochen Kupperschmidt :License: Revised BSD (see `LICENSE` file for details) """ from datetime import timedelta from pathlib import Path # database connection SQLALCHEMY_ECHO = False # Avoid connection errors after database becomes temporarily # unreachable, then becomes reachable again. SQLALCHEMY_ENGINE_OPTIONS = {'pool_pre_ping': True} # Disable Flask-SQLAlchemy's tracking of object modifications. SQLALCHEMY_TRACK_MODIFICATIONS = False # job queue JOBS_ASYNC = True # metrics METRICS_ENABLED = False # RQ dashboard (for job queue) RQ_DASHBOARD_ENABLED = False RQ_DASHBOARD_POLL_INTERVAL = 2500 RQ_DASHBOARD_WEB_BACKGROUND = 'white' # login sessions PERMANENT_SESSION_LIFETIME = timedelta(14) # localization LOCALE = 'de_DE.UTF-8' LOCALES_FORMS = ['de'] TIMEZONE = 'Europe/Berlin' BABEL_DEFAULT_LOCALE = LOCALE BABEL_DEFAULT_TIMEZONE = TIMEZONE # static content files path PATH_DATA = Path('./data') # home page ROOT_REDIRECT_TARGET = None # shop SHOP_ORDER_EXPORT_TIMEZONE = 'Europe/Berlin'
""" byceps.config_defaults ~~~~~~~~~~~~~~~~~~~~~~ Default configuration values :Copyright: 2006-2021 Jochen Kupperschmidt :License: Revised BSD (see `LICENSE` file for details) """ from datetime import timedelta from pathlib import Path # database connection SQLALCHEMY_ECHO = False # Avoid connection errors after database becomes temporarily # unreachable, then becomes reachable again. SQLALCHEMY_ENGINE_OPTIONS = {'pool_pre_ping': True} # Disable Flask-SQLAlchemy's tracking of object modifications. SQLALCHEMY_TRACK_MODIFICATIONS = False # job queue JOBS_ASYNC = True # metrics METRICS_ENABLED = False # RQ dashboard (for job queue) RQ_DASHBOARD_ENABLED = False RQ_DASHBOARD_POLL_INTERVAL = 2500 RQ_DASHBOARD_WEB_BACKGROUND = 'white' # login sessions PERMANENT_SESSION_LIFETIME = timedelta(14) SESSION_COOKIE_SAMESITE = 'Lax' # localization LOCALE = 'de_DE.UTF-8' LOCALES_FORMS = ['de'] TIMEZONE = 'Europe/Berlin' BABEL_DEFAULT_LOCALE = LOCALE BABEL_DEFAULT_TIMEZONE = TIMEZONE # static content files path PATH_DATA = Path('./data') # home page ROOT_REDIRECT_TARGET = None # shop SHOP_ORDER_EXPORT_TIMEZONE = 'Europe/Berlin'
Set session cookie flag `SameSite` to `Lax` (instead of `None`)
Set session cookie flag `SameSite` to `Lax` (instead of `None`)
Python
bsd-3-clause
homeworkprod/byceps,homeworkprod/byceps,homeworkprod/byceps
--- +++ @@ -35,6 +35,7 @@ # login sessions PERMANENT_SESSION_LIFETIME = timedelta(14) +SESSION_COOKIE_SAMESITE = 'Lax' # localization LOCALE = 'de_DE.UTF-8'