commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
5d33de3868df4549621763db07267ef59fb94eb8
dataset/models/tf/losses/core.py
dataset/models/tf/losses/core.py
""" Contains base tf losses """ import tensorflow as tf def softmax_cross_entropy(labels, logits, *args, **kwargs): """ Multi-class CE which takes plain or one-hot labels Parameters ---------- labels : tf.Tensor logits : tf.Tensor args other positional parameters from `tf.losses.softmax_cross_entropy` kwargs other named parameters from `tf.losses.softmax_cross_entropy` Returns ------- tf.Tensor """ labels_shape = tf.shape(labels) logits_shape = tf.shape(logits) c = tf.cast(tf.equal(labels_shape, logits_shape), tf.int32) e = tf.equal(tf.reduce_sum(c, axis=-1), logits_shape.shape[-1]) labels = tf.cond(e, lambda: tf.cast(labels, dtype=logits.dtype), lambda: tf.one_hot(labels, logits_shape[-1], dtype=logits.dtype)) return tf.losses.softmax_cross_entropy(labels, logits, *args, **kwargs)
""" Contains base tf losses """ import tensorflow as tf def softmax_cross_entropy(labels, logits, *args, **kwargs): """ Multi-class CE which takes plain or one-hot labels Parameters ---------- labels : tf.Tensor logits : tf.Tensor args other positional parameters from `tf.losses.softmax_cross_entropy` kwargs other named parameters from `tf.losses.softmax_cross_entropy` Returns ------- tf.Tensor """ labels_shape = tf.shape(labels) logits_shape = tf.shape(logits) c = tf.cast(tf.equal(labels_shape, logits_shape), tf.int32) e = tf.equal(tf.reduce_sum(c, axis=-1), logits_shape.shape[-1]) labels = tf.cond(e, lambda: tf.cast(labels, dtype=logits.dtype), lambda: tf.one_hot(tf.cast(labels, tf.int32), logits_shape[-1], dtype=logits.dtype)) return tf.losses.softmax_cross_entropy(labels, logits, *args, **kwargs)
Fix ohe type in ce
Fix ohe type in ce
Python
apache-2.0
analysiscenter/dataset
--- +++ @@ -25,5 +25,5 @@ c = tf.cast(tf.equal(labels_shape, logits_shape), tf.int32) e = tf.equal(tf.reduce_sum(c, axis=-1), logits_shape.shape[-1]) labels = tf.cond(e, lambda: tf.cast(labels, dtype=logits.dtype), - lambda: tf.one_hot(labels, logits_shape[-1], dtype=logits.dtype)) + lambda: tf.one_hot(tf.cast(labels, tf.int32), logits_shape[-1], dtype=logits.dtype)) return tf.losses.softmax_cross_entropy(labels, logits, *args, **kwargs)
aef238386c71d52def424c8f47a103bd25f12e26
server/proposal/migrations/0034_fix_updated.py
server/proposal/migrations/0034_fix_updated.py
import django.contrib.gis.db.models.fields from django.db import migrations from django.contrib.gis.db.models import Max def fix_updated(apps, _): Proposal = apps.get_model("proposal", "Proposal") proposals = Proposal.objects.annotate(published=Max("documents__published")) for proposal in proposals: if proposal.published: proposal.updated = proposal.published proposal.save() class Migration(migrations.Migration): dependencies = [ ('proposal', '0033_non_null_started'), ] operations = [ migrations.RunPython(fix_updated), ]
import django.contrib.gis.db.models.fields from django.db import migrations from django.contrib.gis.db.models import Max def fix_updated(apps, _): Proposal = apps.get_model("proposal", "Proposal") proposals = Proposal.objects.annotate(published=Max("documents__published")) for proposal in proposals: if proposal.published: proposal.updated = proposal.published proposal.save() def do_nothing(apps, _): pass class Migration(migrations.Migration): dependencies = [ ('proposal', '0033_non_null_started'), ] operations = [ migrations.RunPython(fix_updated, do_nothing), ]
Make fix_updated migration (sort of) reversible
Make fix_updated migration (sort of) reversible
Python
mit
cityofsomerville/citydash,cityofsomerville/citydash,codeforboston/cornerwise,codeforboston/cornerwise,codeforboston/cornerwise,codeforboston/cornerwise,cityofsomerville/citydash,cityofsomerville/cornerwise,cityofsomerville/citydash,cityofsomerville/cornerwise,cityofsomerville/cornerwise,cityofsomerville/cornerwise
--- +++ @@ -12,11 +12,15 @@ proposal.save() +def do_nothing(apps, _): + pass + + class Migration(migrations.Migration): dependencies = [ ('proposal', '0033_non_null_started'), ] operations = [ - migrations.RunPython(fix_updated), + migrations.RunPython(fix_updated, do_nothing), ]
6e526a173de970f2cc8f7cd62823a257786e348e
category/urls.py
category/urls.py
from django.conf.urls import patterns, url from .views import CategoriesList, GalleryDetail, StoryDetail urlpatterns = patterns('category.views', url(r'^gallery/categories/(?P<slug>.+)/$', GalleryDetail.as_view(), name='gallery-detail'), url(r'^stories/categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='story-detail'), url(r'^categories/$', CategoriesList.as_view(), name='category-list'), url(r'^gallery/$', 'gallery_home_page', name='gallery-home-page') )
from django.conf.urls import patterns, url from .views import CategoriesList, GalleryDetail, StoryDetail urlpatterns = patterns('category.views', url(r'^categories/$', CategoriesList.as_view(), name='category-list'), url(r'^categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='category-detail'), url(r'^stories/categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='story-detail'), url(r'^gallery/categories/(?P<slug>.+)/$', GalleryDetail.as_view(), name='gallery-detail'), url(r'^gallery/$', 'gallery_home_page', name='gallery-home-page') )
Add the missing category-detail urlconf to not to break bookmarked users
Add the missing category-detail urlconf to not to break bookmarked users
Python
bsd-3-clause
PARINetwork/pari,PARINetwork/pari,PARINetwork/pari,PARINetwork/pari
--- +++ @@ -3,8 +3,10 @@ from .views import CategoriesList, GalleryDetail, StoryDetail urlpatterns = patterns('category.views', + url(r'^categories/$', CategoriesList.as_view(), name='category-list'), + url(r'^categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='category-detail'), + url(r'^stories/categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='story-detail'), url(r'^gallery/categories/(?P<slug>.+)/$', GalleryDetail.as_view(), name='gallery-detail'), - url(r'^stories/categories/(?P<slug>.+)/$', StoryDetail.as_view(), name='story-detail'), - url(r'^categories/$', CategoriesList.as_view(), name='category-list'), + url(r'^gallery/$', 'gallery_home_page', name='gallery-home-page') )
077ad3e5227c3ad9831a6c94c14cd640f7e933d9
carusele/models.py
carusele/models.py
from django.db import models class News (models.Model): """ News model represent detail description and content of each carusele element. """ title = models.CharField(max_length=400) description = models.TextField(default="") content = models.TextField() pubdate = models.DateTimeField() image = models.ImageField(upload_to="media") def __unicode__(self): return unicode(self.title) def get_absolute_url(self): return "/carusele/art/%i/" % self.id class Element (models.Model): """ This model presents picture and short description of news in carusele javascript element on main page. """ description = models.CharField(max_length=400) image = models.ImageField(upload_to="media") news = models.OneToOneField("News") def __unicode__(self): return unicode(self.description)
from django.core.urlresolvers import reverse from django.db import models class News (models.Model): """ News model represent detail description and content of each carusele element. """ title = models.CharField(max_length=400) description = models.TextField(default="") content = models.TextField() pubdate = models.DateTimeField() image = models.ImageField(upload_to="media") def __unicode__(self): return unicode(self.title) def get_absolute_url(self): return reverse("article", args=(self.id,)) class Element (models.Model): """ This model presents picture and short description of news in carusele javascript element on main page. """ description = models.CharField(max_length=400) image = models.ImageField(upload_to="media") news = models.OneToOneField("News") def __unicode__(self): return unicode(self.description)
Use reverse function for urls in carusele app
Use reverse function for urls in carusele app
Python
apache-2.0
SarFootball/backend,SarFootball/backend,SarFootball/backend
--- +++ @@ -1,3 +1,4 @@ +from django.core.urlresolvers import reverse from django.db import models @@ -16,7 +17,7 @@ return unicode(self.title) def get_absolute_url(self): - return "/carusele/art/%i/" % self.id + return reverse("article", args=(self.id,)) class Element (models.Model):
4aba708916984c61cc7f5fd205d66e8f64634589
main/widgets.py
main/widgets.py
from django_filters.widgets import RangeWidget class CustomRangeWidget(RangeWidget): def __init__(self, widget, attrs=None): attrs_start = {'placeholder': 'От', **(attrs or {})} attrs_stop = {'placeholder': 'До', **(attrs or {})} widgets = (widget(attrs_start), widget(attrs_stop)) super(RangeWidget, self).__init__(widgets, attrs) def format_output(self, rendered_widgets): rendered_widgets.insert(1, ' — ') return '<table class="range-widget"><tr><td>' + '</td><td>'.join(rendered_widgets) + '</td></tr></table>'
from django_filters.widgets import RangeWidget class CustomRangeWidget(RangeWidget): def __init__(self, widget, attrs={}): attrs_start = {'placeholder': 'От'} attrs_start.update(attrs) attrs_stop = {'placeholder': 'До'} attrs_stop.update(attrs) super(RangeWidget, self).__init__((widget(attrs_start), widget(attrs_stop)), attrs) def format_output(self, rendered_widgets): rendered_widgets.insert(1, ' — ') return '<table class="range-widget"><tr><td>' + '</td><td>'.join(rendered_widgets) + '</td></tr></table>'
Make compatible with python 3.4
Make compatible with python 3.4
Python
agpl-3.0
Davidyuk/witcoin,Davidyuk/witcoin
--- +++ @@ -2,11 +2,12 @@ class CustomRangeWidget(RangeWidget): - def __init__(self, widget, attrs=None): - attrs_start = {'placeholder': 'От', **(attrs or {})} - attrs_stop = {'placeholder': 'До', **(attrs or {})} - widgets = (widget(attrs_start), widget(attrs_stop)) - super(RangeWidget, self).__init__(widgets, attrs) + def __init__(self, widget, attrs={}): + attrs_start = {'placeholder': 'От'} + attrs_start.update(attrs) + attrs_stop = {'placeholder': 'До'} + attrs_stop.update(attrs) + super(RangeWidget, self).__init__((widget(attrs_start), widget(attrs_stop)), attrs) def format_output(self, rendered_widgets): rendered_widgets.insert(1, ' — ')
94d66121368906b52fa8a9f214813b7b798c2b5b
lib/custom_data/settings_manager.py
lib/custom_data/settings_manager.py
"""This module provides functions for saving to and loading data from the settings XML file. Attributes: SETTINGS_PATH Filepath for the settings file. """ SETTINGS_PATH = 'settings.xml'
"""This module provides functions for saving to and loading data from the settings XML file. Attributes: SETTINGS_PATH (String): The file path for the settings file. SETTINGS_SCHEMA_PATH (String): The file path for the settings' XML Schema. """ SETTINGS_PATH = 'settings.xml' SETTINGS_SCHEMA_PATH = 'settings.xsd'
Add constant for settings schema file path
Add constant for settings schema file path
Python
unlicense
MarquisLP/Sidewalk-Champion
--- +++ @@ -2,7 +2,9 @@ the settings XML file. Attributes: - SETTINGS_PATH Filepath for the settings file. + SETTINGS_PATH (String): The file path for the settings file. + SETTINGS_SCHEMA_PATH (String): The file path for the settings' + XML Schema. """ - SETTINGS_PATH = 'settings.xml' +SETTINGS_SCHEMA_PATH = 'settings.xsd'
60ea2738b39b38bdc1f25594a759aace0f520501
web/manage.py
web/manage.py
from flask.ext.script import Manager from app import get_app from create_db import drop_all_tables, create_barebones_data, create_all_data, create_server app = get_app('config.DockerConfiguration') manager = Manager(app) manager.command(drop_all_tables) manager.command(create_barebones_data) manager.command(create_all_data) manager.command(create_server) if __name__ == '__main__': manager.run()
from flask.ext.script import Manager from app import get_app from create_db import drop_all_tables, create_barebones_data, create_all_data, create_server app = get_app('config.DockerConfiguration') manager = Manager(app) manager.command(drop_all_tables) manager.command(create_barebones_data) manager.command(create_all_data) manager.command(create_server) @manager.command def config(): 'Print out all config values from the fully assembled flask app' print('\n'.join('%s=%s' % item for item in sorted(app.config.items()))) if __name__ == '__main__': manager.run()
Add utility function to dump flask env
Add utility function to dump flask env
Python
mit
usgo/online-ratings,usgo/online-ratings,usgo/online-ratings,Kashomon/online-ratings,Kashomon/online-ratings,Kashomon/online-ratings
--- +++ @@ -12,6 +12,10 @@ manager.command(create_all_data) manager.command(create_server) +@manager.command +def config(): + 'Print out all config values from the fully assembled flask app' + print('\n'.join('%s=%s' % item for item in sorted(app.config.items()))) if __name__ == '__main__': manager.run()
39946f9fa5127d240d7147d50b676ad083514e85
campus02/urls.py
campus02/urls.py
#!/usr/bin/python # -*- coding: utf-8 -*- from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns( '', url(r'^', include('django.contrib.auth.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^web/', include('campus02.web.urls', namespace='web')), url(r'^', include('campus02.base.urls', namespace='base')), )
#!/usr/bin/python # -*- coding: utf-8 -*- from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib import admin urlpatterns = patterns( '', url(r'^', include('django.contrib.auth.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^web/', include('campus02.web.urls', namespace='web')), url(r'^', include('campus02.base.urls', namespace='base')), ) if settings.DEBUG: import debug_toolbar urlpatterns += patterns( '', url(r'^__debug__/', include(debug_toolbar.urls)), )
Add custom debug toolbar URL mount point.
Add custom debug toolbar URL mount point.
Python
mit
fladi/django-campus02,fladi/django-campus02
--- +++ @@ -1,6 +1,7 @@ #!/usr/bin/python # -*- coding: utf-8 -*- +from django.conf import settings from django.conf.urls import patterns, include, url from django.contrib import admin @@ -11,3 +12,10 @@ url(r'^web/', include('campus02.web.urls', namespace='web')), url(r'^', include('campus02.base.urls', namespace='base')), ) + +if settings.DEBUG: + import debug_toolbar + urlpatterns += patterns( + '', + url(r'^__debug__/', include(debug_toolbar.urls)), + )
ce990bfb3c742c9f19f0af43a10aad8193fa084c
keystoneclient_kerberos/__init__.py
keystoneclient_kerberos/__init__.py
# -*- coding: utf-8 -*- # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import pbr.version from keystoneclient_kerberos import v3 __version__ = pbr.version.VersionInfo( 'keystoneclient_kerberos').version_string() V3Kerberos = v3.Kerberos __all__ = ['V3Kerberos']
# -*- coding: utf-8 -*- # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import pbr.version from keystoneclient_kerberos import v3 __version__ = pbr.version.VersionInfo( 'python-keystoneclient-kerberos').version_string() V3Kerberos = v3.Kerberos __all__ = ['V3Kerberos']
Use the package name when looking up version
Use the package name when looking up version We need to give PBR the full package name when it looks up the version from a pip installed package. keystoneclient_kerberos is the module name, not the package name. Change-Id: I638d0975d77db3767c3675dceb05466388abebc9 Closes-Bug: #1441918
Python
apache-2.0
cernops/python-keystoneclient-kerberos
--- +++ @@ -17,7 +17,7 @@ from keystoneclient_kerberos import v3 __version__ = pbr.version.VersionInfo( - 'keystoneclient_kerberos').version_string() + 'python-keystoneclient-kerberos').version_string() V3Kerberos = v3.Kerberos
433d9b2c1c29f32a7d5289e84673308c96302d8d
controlers/access.py
controlers/access.py
'''Copyright(C): Leaf Johnson 2011 This file is part of makeclub. makeclub is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. makeclub is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU Affero General Public License along with makeclub. If not, see <http://www.gnu.org/licenses/>. ''' from google.appengine.api import users operations = [ "listclubs", "createClub" ] clubOperations = [ "view", "create", "edit", "delete", "arrange", "finish" ] actOperatoins = [ "create", "view", "edit", "join" ] def isAccessible (user, operation): return True def hasClubPrivilige (user, club, operation): return True def hasClubPrivilige (user, act, operation): return True
'''Copyright(C): Leaf Johnson 2011 This file is part of makeclub. makeclub is free software: you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation, either version 3 of the License, or (at your option) any later version. makeclub is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU Affero General Public License along with makeclub. If not, see <http://www.gnu.org/licenses/>. ''' from google.appengine.api import users operations = [ "listclubs", "createClub" ] clubOperations = [ "view", "create", "edit", "delete", "arrange", "finish", "newact" ] actOperatoins = [ "view", "edit", "join" ] def isAccessible (user, operation): return True def hasClubPrivilige (user, club, operation): return True def hasActPrivilige (user, act, operation): return True
FIX a bug, you fuck'in forgot to rename the new function
FIX a bug, you fuck'in forgot to rename the new function
Python
agpl-3.0
cardmaster/makeclub,cardmaster/makeclub,cardmaster/makeclub
--- +++ @@ -31,11 +31,11 @@ "edit", "delete", "arrange", - "finish" + "finish", + "newact" ] actOperatoins = [ - "create", "view", "edit", "join" @@ -47,6 +47,6 @@ def hasClubPrivilige (user, club, operation): return True -def hasClubPrivilige (user, act, operation): +def hasActPrivilige (user, act, operation): return True
addc1e83911f72282eca9603e2c483ba6ef5ef7c
packages/xsp.py
packages/xsp.py
GitHubTarballPackage('mono', 'xsp', '2.11', 'd3e2f80ff59ddff68e757a520655555e2fbf2695', configure = './autogen.sh --prefix="%{prefix}"')
GitHubTarballPackage('mono', 'xsp', '3.0.11', '4587438369691b9b3e8415e1f113aa98b57d1fde', configure = './autogen.sh --prefix="%{prefix}"')
Update to the latest XSP.
Update to the latest XSP.
Python
mit
BansheeMediaPlayer/bockbuild,mono/bockbuild,BansheeMediaPlayer/bockbuild,BansheeMediaPlayer/bockbuild,mono/bockbuild
--- +++ @@ -1 +1 @@ -GitHubTarballPackage('mono', 'xsp', '2.11', 'd3e2f80ff59ddff68e757a520655555e2fbf2695', configure = './autogen.sh --prefix="%{prefix}"') +GitHubTarballPackage('mono', 'xsp', '3.0.11', '4587438369691b9b3e8415e1f113aa98b57d1fde', configure = './autogen.sh --prefix="%{prefix}"')
d8b29fd094a7a2d74c74e32b05a810930655fb47
src/modules/phython.py
src/modules/phython.py
import json import runpy import sys def run(): args = sys.argv if len(args) < 3: raise Exception('Both module name and function name are required.') module, function = args[1:3] module = runpy.run_module(module) if function not in module: raise Exception(function + ' is not defined in ' + module['__file__']) call(module[function]) def call(function): arguments = raw_input().strip() arguments = json.loads(arguments) output = function(*arguments) print(json.dumps(output)) run()
import json import runpy import sys def run(): args = sys.argv if len(args) < 3: raise Exception('Both module name and function name are required.') module, function = args[1:3] module = runpy.run_module(module) if function not in module: raise Exception(function + ' is not defined in ' + module['__file__']) call(module[function]) def call(function): # raw_input() is removed in python 3 try: input = raw_input except NameError: pass arguments = input().strip() arguments = json.loads(arguments) output = function(*arguments) print(json.dumps(output)) run()
Fix raw_input() error in python 3
Fix raw_input() error in python 3
Python
mit
marella/phython,marella/phython,marella/phython
--- +++ @@ -17,7 +17,12 @@ call(module[function]) def call(function): - arguments = raw_input().strip() + # raw_input() is removed in python 3 + try: + input = raw_input + except NameError: + pass + arguments = input().strip() arguments = json.loads(arguments) output = function(*arguments) print(json.dumps(output))
bc1d19800d58291f4c4392d041a7913602fe8c7d
dallinger/jupyter.py
dallinger/jupyter.py
from ipywidgets import widgets from jinja2 import Template from traitlets import ( observe, Unicode, ) from dallinger.config import get_config header_template = Template(u""" <h2>{{ name }}</h2> <div>Status: {{ status }}</div> {% if app_id %}<div>App ID: {{ app_id }}</div>{% endif %} """) config_template = Template(u""" <table style="min-width: 50%"> {% for k, v in config %} <tr> <th>{{ k }}</th> <td>{{ v }}</td> </tr> {% endfor %} </table> """) class ExperimentWidget(widgets.VBox): status = Unicode('Unknown') def __init__(self, exp): self.exp = exp super(ExperimentWidget, self).__init__() self.render() @observe('status') def render(self, change=None): header = widgets.HTML( header_template.render( name=self.exp.task, status=self.status, app_id=self.exp.app_id, ), ) config = get_config() if config.ready: config_items = config.as_dict().items() config_items.sort() config_tab = widgets.HTML( config_template.render(config=config_items) ) else: config_tab = widgets.HTML('Not loaded.') tabs = widgets.Tab(children=[config_tab]) tabs.set_title(0, 'Configuration') self.children = [header, tabs]
from ipywidgets import widgets from jinja2 import Template from traitlets import ( observe, Unicode, ) from dallinger.config import get_config header_template = Template(u""" <h2>{{ name }}</h2> <div>Status: {{ status }}</div> {% if app_id %}<div>App ID: {{ app_id }}</div>{% endif %} """) config_template = Template(u""" <table style="min-width: 50%"> {% for k, v in config %} <tr> <th>{{ k }}</th> <td>{{ v }}</td> </tr> {% endfor %} </table> """) class ExperimentWidget(widgets.VBox): status = Unicode('Unknown') def __init__(self, exp): self.exp = exp super(ExperimentWidget, self).__init__() self.render() @observe('status') def render(self, change=None): header = widgets.HTML( header_template.render( name=self.exp.task, status=self.status, app_id=self.exp.app_id, ), ) config = get_config() if config.ready: config_items = list(config.as_dict().items()) config_items.sort() config_tab = widgets.HTML( config_template.render(config=config_items) ) else: config_tab = widgets.HTML('Not loaded.') tabs = widgets.Tab(children=[config_tab]) tabs.set_title(0, 'Configuration') self.children = [header, tabs]
Fix sorting dict items in python 3
Fix sorting dict items in python 3
Python
mit
jcpeterson/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger
--- +++ @@ -45,7 +45,7 @@ ) config = get_config() if config.ready: - config_items = config.as_dict().items() + config_items = list(config.as_dict().items()) config_items.sort() config_tab = widgets.HTML( config_template.render(config=config_items)
745c9445e16f72dbc1791abef2b7f52eb5e1f093
open_spiel/python/tests/referee_test.py
open_spiel/python/tests/referee_test.py
# Copyright 2019 DeepMind Technologies Ltd. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for open_spiel.python.referee.""" import os import pyspiel from absl.testing import absltest class RefereeTest(absltest.TestCase): def test_playing_tournament(self): base = os.path.dirname(__file__) + "/../../higc/bots" ref = pyspiel.Referee("kuhn_poker", [f"{base}/random_bot_py.sh", f"{base}/random_bot_cpp.sh"]) results = ref.play_tournament(num_matches=1) self.assertEqual(len(results.matches), 1) if __name__ == "__main__": absltest.main()
# Copyright 2019 DeepMind Technologies Ltd. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for open_spiel.python.referee.""" import os import pyspiel from absl.testing import absltest class RefereeTest(absltest.TestCase): def test_playing_tournament(self): base = os.path.dirname(__file__) + "/../../higc/bots" ref = pyspiel.Referee( "kuhn_poker", [f"{base}/random_bot_py.sh", f"{base}/random_bot_cpp.sh"], settings=pyspiel.TournamentSettings(timeout_ready=2000, timeout_start=2000) ) results = ref.play_tournament(num_matches=1) self.assertEqual(len(results.matches), 1) if __name__ == "__main__": absltest.main()
Increase timeouts for the python test.
Increase timeouts for the python test.
Python
apache-2.0
deepmind/open_spiel,deepmind/open_spiel,deepmind/open_spiel,deepmind/open_spiel,deepmind/open_spiel,deepmind/open_spiel
--- +++ @@ -22,9 +22,12 @@ class RefereeTest(absltest.TestCase): def test_playing_tournament(self): base = os.path.dirname(__file__) + "/../../higc/bots" - ref = pyspiel.Referee("kuhn_poker", - [f"{base}/random_bot_py.sh", - f"{base}/random_bot_cpp.sh"]) + ref = pyspiel.Referee( + "kuhn_poker", + [f"{base}/random_bot_py.sh", f"{base}/random_bot_cpp.sh"], + settings=pyspiel.TournamentSettings(timeout_ready=2000, + timeout_start=2000) + ) results = ref.play_tournament(num_matches=1) self.assertEqual(len(results.matches), 1)
e912c76ec60abf9a8263e65d8df8d466518f57b2
pysswords/db.py
pysswords/db.py
from glob import glob import os import shutil from .credential import Credential from .crypt import create_gpg, load_gpg class Database(object): def __init__(self, path, gpg): self.path = path self.gpg = gpg @classmethod def create(cls, path, passphrase, gpg_bin="gpg"): gpg = create_gpg(gpg_bin, path, passphrase) return Database(path, gpg) @classmethod def from_path(cls, path, gpg_bin="gpg"): gpg = load_gpg(binary=gpg_bin, database_path=path) return Database(path, gpg) @property def gpg_key(self): return self.gpg.list_keys(secret=True)[0]["fingerprint"] @property def credentials(self): return [self.credential(os.path.basename(c)) for c in glob(self.path + "/**")] def add(self, credential): encrypted_password = self.gpg.encrypt( credential.password, self.gpg_key ) credential.password = str(encrypted_password) credential.save(database_path=self.path) def delete(self, name): credential_path = os.path.join(self.path, name) shutil.rmtree(credential_path) def credential(self, name): credential_path = os.path.join(self.path, name) credential = Credential.from_path(credential_path) return credential def search(self, query): return [c for c in self.credentials if query in str(c)]
from glob import glob import os import shutil from .credential import Credential from .crypt import create_gpg, load_gpg class Database(object): def __init__(self, path, gpg): self.path = path self.gpg = gpg @classmethod def create(cls, path, passphrase, gpg_bin="gpg"): gpg = create_gpg(gpg_bin, path, passphrase) return Database(path, gpg) @classmethod def from_path(cls, path, gpg_bin="gpg"): gpg = load_gpg(binary=gpg_bin, database_path=path) return Database(path, gpg) @property def gpg_key(self): return self.gpg.list_keys(secret=True)[0]["fingerprint"] @property def credentials(self): return [self.credential(os.path.basename(c)) for c in glob(self.path + "/**")] def add(self, credential): encrypted_password = self.gpg.encrypt( credential.password, self.gpg_key, cipher_algo="AES256" ) credential.password = str(encrypted_password) credential.save(database_path=self.path) def delete(self, name): credential_path = os.path.join(self.path, name) shutil.rmtree(credential_path) def credential(self, name): credential_path = os.path.join(self.path, name) credential = Credential.from_path(credential_path) return credential def search(self, query): return [c for c in self.credentials if query in str(c)]
Change GPG password encryption to AES256
Change GPG password encryption to AES256
Python
mit
scorphus/passpie,eiginn/passpie,marcwebbie/passpie,marcwebbie/pysswords,scorphus/passpie,eiginn/passpie,marcwebbie/passpie
--- +++ @@ -34,7 +34,8 @@ def add(self, credential): encrypted_password = self.gpg.encrypt( credential.password, - self.gpg_key + self.gpg_key, + cipher_algo="AES256" ) credential.password = str(encrypted_password) credential.save(database_path=self.path)
57d008222a531ec79957611dc077a67499426986
CS480/milestone4/driver.py
CS480/milestone4/driver.py
#!/usr/bin/env python from sys import * from myreglexer import * from getopt import * from myparser import * from tree import * from semantic import * contents = [] def main(argv): try: opts, args = getopt(argv, "h", ["help"]) except GetoptError: usage() sys.exit(1) for opt, arg in opts: if opt in ("-h", "--help"): usage() sys.exit() for arg in args: try: with open(arg, 'r') as f: readin = f.read() tree = parser(readin) #tree.traverse_pre() temp = tree.build_stack_pre() #print temp print semantic_check(temp) temp = tree.build_stack_post() for i in temp: print i.data[1] #tree.traverse_post() except IOError: print "File %s not found!" % arg sys.exit(1) def no_args(): for line in sys.stdin: contents.append(line) def usage(): print "Usage: driver.py [-h] infile1 infile2 infile3 ..." print "Or just take intput from stdin" if __name__ == "__main__": if len(sys.argv) < 2: no_args() #usage() #sys.exit(1) main(sys.argv[1:])
#!/usr/bin/env python from sys import * from myreglexer import * from getopt import * from myparser import * from tree import * from semantic import * contents = [] def main(argv): try: opts, args = getopt(argv, "h", ["help"]) except GetoptError: usage() sys.exit(1) for opt, arg in opts: if opt in ("-h", "--help"): usage() sys.exit() for arg in args: try: with open(arg, 'r') as f: readin = f.read() tree = parser(readin) #tree.traverse_pre() temp = tree.build_stack_pre() #print temp semantic_check(temp) temp = tree.build_stack_post() for i in temp: print i.data[1] #tree.traverse_post() except IOError: print "File %s not found!" % arg sys.exit(1) def no_args(): for line in sys.stdin: contents.append(line) def usage(): print "Usage: driver.py [-h] infile1 infile2 infile3 ..." print "Or just take intput from stdin" if __name__ == "__main__": if len(sys.argv) < 2: no_args() #usage() #sys.exit(1) main(sys.argv[1:])
Change the way semantic is called to allow for looping and handling
Change the way semantic is called to allow for looping and handling [SS]
Python
apache-2.0
stumped2/school,stumped2/school,stumped2/school,stumped2/school,stumped2/school,stumped2/school,stumped2/school,stumped2/school
--- +++ @@ -29,7 +29,7 @@ #tree.traverse_pre() temp = tree.build_stack_pre() #print temp - print semantic_check(temp) + semantic_check(temp) temp = tree.build_stack_post() for i in temp: print i.data[1]
98cbd5207bd25fb0fafd25f18870c771479255e1
run-tests.py
run-tests.py
#!/usr/bin/env python3 import os import subprocess import sys args = [ sys.executable or 'python', # Python interpreter to call for testing. '-B', # Don't write .pyc files on import. '-m', 'unittest', # Run the unittest module as a script. 'discover', # Use test discovery. '-s', 'tests', # Start discovery in 'tests' directory. '-t', os.path.dirname(__file__), # Set top-level of project. ] args.extend(sys.argv[1:]) # Append any arguments passed to script (-v, etc.). sys.exit(subprocess.call(args))
#!/usr/bin/env python3 import os import subprocess import sys args = [ sys.executable or 'python', # Python interpreter to call for testing. '-B', # Don't write .pyc files on import. '-W', 'default', # Enable default handling for all warnings. '-m', 'unittest', # Run the unittest module as a script. 'discover', # Use test discovery. '-s', 'tests', # Start discovery in 'tests' directory. '-t', os.path.dirname(__file__), # Set top-level of project. ] args.extend(sys.argv[1:]) # Append any arguments passed to script (-v, etc.). sys.exit(subprocess.call(args))
Enable default warnings while testing.
Enable default warnings while testing.
Python
mit
shawnbrown/gpn,shawnbrown/gpn
--- +++ @@ -7,6 +7,7 @@ args = [ sys.executable or 'python', # Python interpreter to call for testing. '-B', # Don't write .pyc files on import. + '-W', 'default', # Enable default handling for all warnings. '-m', 'unittest', # Run the unittest module as a script. 'discover', # Use test discovery. '-s', 'tests', # Start discovery in 'tests' directory.
df155919eb81748231b5b7f834e0739d78a38471
tests/CLI/modules/subnet_tests.py
tests/CLI/modules/subnet_tests.py
""" SoftLayer.tests.CLI.modules.subnet_tests ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :license: MIT, see LICENSE for more details. """ from SoftLayer import testing import json class SubnetTests(testing.TestCase): def test_detail(self): result = self.run_command(['subnet', 'detail', '1234']) self.assert_no_fail(result) self.assertEqual( { 'id': 1234, 'identifier': '1.2.3.4/26', 'subnet type': 'ADDITIONAL_PRIMARY', 'network space': 'PUBLIC', 'gateway': '1.2.3.254', 'broadcast': '1.2.3.255', 'datacenter': 'dal10', 'vs': [ { 'hostname': 'hostname0', 'domain': 'sl.test', 'public_ip': '1.2.3.10', 'private_ip': '10.0.1.2' } ], 'hardware': 'none', 'usable ips': 22 }, json.loads(result.output))
""" SoftLayer.tests.CLI.modules.subnet_tests ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ :license: MIT, see LICENSE for more details. """ from SoftLayer import testing import json class SubnetTests(testing.TestCase): def test_detail(self): result = self.run_command(['subnet', 'detail', '1234']) self.assert_no_fail(result) self.assertEqual( { 'id': 1234, 'identifier': '1.2.3.4/26', 'subnet type': 'ADDITIONAL_PRIMARY', 'network space': 'PUBLIC', 'gateway': '1.2.3.254', 'broadcast': '1.2.3.255', 'datacenter': 'dal10', 'vs': [ { 'hostname': 'hostname0', 'domain': 'sl.test', 'public_ip': '1.2.3.10', 'private_ip': '10.0.1.2' } ], 'hardware': 'none', 'usable ips': 22 }, json.loads(result.output))
Fix style nit, line end for test file
Fix style nit, line end for test file
Python
mit
allmightyspiff/softlayer-python,softlayer/softlayer-python,nanjj/softlayer-python,kyubifire/softlayer-python
d6e4aa32b7b79adc734dfc2b058509cedf771944
munigeo/migrations/0004_building.py
munigeo/migrations/0004_building.py
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2016-11-10 08:46 from __future__ import unicode_literals import django.contrib.gis.db.models.fields from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('munigeo', '0003_add_modified_time_to_address_and_street'), ] operations = [ migrations.CreateModel( name='Building', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('origin_id', models.CharField(db_index=True, max_length=40)), ('geometry', django.contrib.gis.db.models.fields.MultiPolygonField(srid=4326)), ('modified_at', models.DateTimeField(auto_now=True, help_text='Time when the information was last changed')), ('addresses', models.ManyToManyField(blank=True, to='munigeo.Address')), ('municipality', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='munigeo.Municipality')), ], options={ 'ordering': ['municipality', 'origin_id'], }, ), ]
# -*- coding: utf-8 -*- # Generated by Django 1.10.1 on 2016-11-10 08:46 from __future__ import unicode_literals import django.contrib.gis.db.models.fields from django.db import migrations, models import django.db.models.deletion from munigeo.utils import get_default_srid DEFAULT_SRID = get_default_srid() class Migration(migrations.Migration): dependencies = [ ('munigeo', '0003_add_modified_time_to_address_and_street'), ] operations = [ migrations.CreateModel( name='Building', fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('origin_id', models.CharField(db_index=True, max_length=40)), ('geometry', django.contrib.gis.db.models.fields.MultiPolygonField(srid=DEFAULT_SRID)), ('modified_at', models.DateTimeField(auto_now=True, help_text='Time when the information was last changed')), ('addresses', models.ManyToManyField(blank=True, to='munigeo.Address')), ('municipality', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='munigeo.Municipality')), ], options={ 'ordering': ['municipality', 'origin_id'], }, ), ]
Fix building migration SRID logic
Fix building migration SRID logic Hardcoding the SRID in the migration could result in a mismatch between the model srid and the migration srid.
Python
agpl-3.0
City-of-Helsinki/munigeo
--- +++ @@ -5,6 +5,9 @@ import django.contrib.gis.db.models.fields from django.db import migrations, models import django.db.models.deletion + +from munigeo.utils import get_default_srid +DEFAULT_SRID = get_default_srid() class Migration(migrations.Migration): @@ -19,7 +22,7 @@ fields=[ ('id', models.AutoField(auto_created=True, primary_key=True, serialize=False, verbose_name='ID')), ('origin_id', models.CharField(db_index=True, max_length=40)), - ('geometry', django.contrib.gis.db.models.fields.MultiPolygonField(srid=4326)), + ('geometry', django.contrib.gis.db.models.fields.MultiPolygonField(srid=DEFAULT_SRID)), ('modified_at', models.DateTimeField(auto_now=True, help_text='Time when the information was last changed')), ('addresses', models.ManyToManyField(blank=True, to='munigeo.Address')), ('municipality', models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, to='munigeo.Municipality')),
35397c33f1b52f158c11941e17211eb699836003
tests/integration/indexer-test.py
tests/integration/indexer-test.py
# -*- coding: utf-8 -*- from nose import tools as nose import unittest from shiva.app import app from shiva.indexer import Indexer class IndexerTestCase(unittest.TestCase): def test_main(self): with app.app_context(): lola = Indexer(app.config) nose.eq_(lola.run(), None)
# -*- coding: utf-8 -*- from nose import tools as nose import unittest from shiva.app import app, db from shiva.indexer import Indexer class IndexerTestCase(unittest.TestCase): def setUp(self): db.create_all() def test_main(self): with app.app_context(): app.config['MEDIA_DIRS'] = [] lola = Indexer(app.config) nose.eq_(lola.run(), None) def tearDown(self): db.drop_all()
Fix to indexer integration tests
Fix to indexer integration tests
Python
mit
tooxie/shiva-server,maurodelazeri/shiva-server,tooxie/shiva-server,maurodelazeri/shiva-server
--- +++ @@ -2,13 +2,20 @@ from nose import tools as nose import unittest -from shiva.app import app +from shiva.app import app, db from shiva.indexer import Indexer class IndexerTestCase(unittest.TestCase): + def setUp(self): + db.create_all() + def test_main(self): with app.app_context(): + app.config['MEDIA_DIRS'] = [] lola = Indexer(app.config) nose.eq_(lola.run(), None) + + def tearDown(self): + db.drop_all()
6421543ff423fc110cd660850f55f7097db5805d
contrib/performance/setbackend.py
contrib/performance/setbackend.py
## # Copyright (c) 2010 Apple Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ## import sys from xml.etree import ElementTree from xml.etree import ElementPath def main(): conf = ElementTree.parse(file(sys.argv[1])) if sys.argv[2] == 'postgresql': value = 'true' elif sys.argv[2] == 'filesystem': value = 'false' else: raise RuntimeError("Don't know what to do with %r" % (sys.argv[2],)) replace(conf.getiterator(), value) conf.write(sys.stdout) def replace(elements, value): found = False for ele in elements: if found: ele.tag = value return if ele.tag == 'key' and ele.text == 'UseDatabase': found = True raise RuntimeError("Failed to find <key>UseDatabase</key>")
## # Copyright (c) 2010 Apple Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ## """ Generate a new calendar server configuration file based on an existing one, with a few values changed to satisfy requirements of the benchmarking tools. """ import sys from xml.etree import ElementTree from xml.etree import ElementPath def main(): conf = ElementTree.parse(file(sys.argv[1])) if sys.argv[2] == 'postgresql': value = 'true' elif sys.argv[2] == 'filesystem': value = 'false' else: raise RuntimeError("Don't know what to do with %r" % (sys.argv[2],)) # Here are the config changes we make - use the specified backend replace(conf.getiterator(), 'UseDatabase', value) # - and disable the response cache replace(conf.getiterator(), 'EnableResponseCache', 'false') conf.write(sys.stdout) def replace(elements, key, value): found = False for ele in elements: if found: ele.tag = value return if ele.tag == 'key' and ele.text == key: found = True raise RuntimeError("Failed to find <key>UseDatabase</key>")
Rewrite the config to disable the response cache, too.
Rewrite the config to disable the response cache, too. git-svn-id: 81e381228600e5752b80483efd2b45b26c451ea2@6929 e27351fd-9f3e-4f54-a53b-843176b1656c
Python
apache-2.0
trevor/calendarserver,trevor/calendarserver,trevor/calendarserver
--- +++ @@ -14,6 +14,12 @@ # limitations under the License. ## +""" +Generate a new calendar server configuration file based on an existing +one, with a few values changed to satisfy requirements of the +benchmarking tools. +""" + import sys from xml.etree import ElementTree from xml.etree import ElementPath @@ -26,16 +32,20 @@ value = 'false' else: raise RuntimeError("Don't know what to do with %r" % (sys.argv[2],)) - replace(conf.getiterator(), value) + + # Here are the config changes we make - use the specified backend + replace(conf.getiterator(), 'UseDatabase', value) + # - and disable the response cache + replace(conf.getiterator(), 'EnableResponseCache', 'false') conf.write(sys.stdout) -def replace(elements, value): +def replace(elements, key, value): found = False for ele in elements: if found: ele.tag = value return - if ele.tag == 'key' and ele.text == 'UseDatabase': + if ele.tag == 'key' and ele.text == key: found = True raise RuntimeError("Failed to find <key>UseDatabase</key>")
0ef0f546c23754cd339adbc00cb3d90558af744c
examples/list_sysdig_captures.py
examples/list_sysdig_captures.py
#!/usr/bin/env python # # Print the list of sysdig captures. # import os import sys sys.path.insert(0, os.path.join(os.path.dirname(os.path.realpath(sys.argv[0])), '..')) from sdcclient import SdcClient # # Parse arguments # if len(sys.argv) != 2: print('usage: %s <sysdig-token>' % sys.argv[0]) print('You can find your token at https://app.sysdigcloud.com/#/settings/user') sys.exit(1) sdc_token = sys.argv[1] # # Instantiate the SDC client # sdclient = SdcClient(sdc_token) # # Fire the request. # ok, res = sdclient.get_sysdig_captures() # # Show the list of metrics # if ok: captures = res[1]['dumps'] else: print(res) sys.exit(1) for capture in captures: print("Folder %s, Name %s, Host: %s, Size: %d, Status: %s" % (capture['folder'], capture['name'], capture['agent']['hostName'], capture['size'], capture['status']))
#!/usr/bin/env python # # Print the list of sysdig captures. # import os import sys sys.path.insert(0, os.path.join(os.path.dirname(os.path.realpath(sys.argv[0])), '..')) from sdcclient import SdcClient # # Parse arguments # if len(sys.argv) != 2: print('usage: %s <sysdig-token>' % sys.argv[0]) print('You can find your token at https://app.sysdigcloud.com/#/settings/user') sys.exit(1) sdc_token = sys.argv[1] # # Instantiate the SDC client # sdclient = SdcClient(sdc_token) # # Fire the request. # ok, res = sdclient.get_sysdig_captures() # # Show the list of metrics # if ok: captures = res['dumps'] else: print(res) sys.exit(1) for capture in captures: print("Folder %s, Name %s, Host: %s, Size: %d, Status: %s" % (capture['folder'], capture['name'], capture['agent']['hostName'], capture['size'], capture['status']))
Fix legacy use of action result
Fix legacy use of action result
Python
mit
draios/python-sdc-client,draios/python-sdc-client
--- +++ @@ -32,7 +32,7 @@ # Show the list of metrics # if ok: - captures = res[1]['dumps'] + captures = res['dumps'] else: print(res) sys.exit(1)
73949126f9c50669da8687b9fae5b8c7db0a89f6
coffee/deamon.py
coffee/deamon.py
#!/home/pi/coffee/venv/bin/python import os import sys import time from coffee.models import Status sys.path.append(os.path.dirname(os.path.dirname(__file__))) DEBUG = 1 PIN = 14 def main(): import RPi.GPIO as GPIO GPIO.setmode(GPIO.BCM) status = Status() def rc_time(RCpin): reading = 0 GPIO.setup(RCpin, GPIO.OUT) GPIO.output(RCpin, GPIO.LOW) time.sleep(0.1) GPIO.setup(RCpin, GPIO.IN) while (GPIO.input(RCpin) == GPIO.LOW): reading += 1 if reading > 5000: return reading return reading while True: if rc_time(PIN) <= 5000: status.update(True) else: status.update(False) main()
#!/home/pi/coffee/venv/bin/python import os import sys import time from coffee.models import Status sys.path.append(os.path.dirname(os.path.dirname(__file__))) DEBUG = 1 # The GPIO pin the button is connected to BUTTON_PIN = 7 # The GPIO pin the button's LED is connected to LED_PIN = 4 def main(): import RPi.GPIO as GPIO status = Status() GPIO.setmode(GPIO.BCM) GPIO.setup(BUTTON_PIN, GPIO.IN, pull_up_down=GPIO.PUD_UP) GPIO.setup(LED_PIN, GPIO.OUT, initial=1) # Blink the LED, leave off def blink_led(num_blinks, blink_duration, blink_pause): for b in range(num_blinks): GPIO.output(LED_PIN, 1) time.sleep(blink_duration) GPIO.output(LED_PIN, 0) time.sleep(blink_pause) GPIO.output(LED_PIN, 0) # Listen for button presses while True: input_value = GPIO.input(BUTTON_PIN) if input_value == False: status.update(True) blink_led(3, 0.3, 0.2) status.update(False) time.sleep(0.1) main()
Write pi code to use button
Write pi code to use button
Python
mit
webkom/coffee,webkom/coffee
--- +++ @@ -10,32 +10,38 @@ DEBUG = 1 -PIN = 14 + +# The GPIO pin the button is connected to +BUTTON_PIN = 7 +# The GPIO pin the button's LED is connected to +LED_PIN = 4 def main(): import RPi.GPIO as GPIO - GPIO.setmode(GPIO.BCM) status = Status() - def rc_time(RCpin): - reading = 0 - GPIO.setup(RCpin, GPIO.OUT) - GPIO.output(RCpin, GPIO.LOW) + GPIO.setmode(GPIO.BCM) + GPIO.setup(BUTTON_PIN, GPIO.IN, pull_up_down=GPIO.PUD_UP) + GPIO.setup(LED_PIN, GPIO.OUT, initial=1) + + # Blink the LED, leave off + def blink_led(num_blinks, blink_duration, blink_pause): + for b in range(num_blinks): + GPIO.output(LED_PIN, 1) + time.sleep(blink_duration) + GPIO.output(LED_PIN, 0) + time.sleep(blink_pause) + GPIO.output(LED_PIN, 0) + + # Listen for button presses + while True: + input_value = GPIO.input(BUTTON_PIN) + if input_value == False: + status.update(True) + blink_led(3, 0.3, 0.2) + status.update(False) time.sleep(0.1) - GPIO.setup(RCpin, GPIO.IN) - while (GPIO.input(RCpin) == GPIO.LOW): - reading += 1 - if reading > 5000: - return reading - return reading - - while True: - if rc_time(PIN) <= 5000: - status.update(True) - else: - status.update(False) - main()
eb46bc61a05279d338c9e1062988f7db67f060fb
makesty.py
makesty.py
import re # Input file created from http://astronautweb.co/snippet/font-awesome/ INPUT_FILE = 'htmlfontawesome.txt' with open(INPUT_FILE) as r: for line in r: # Expects to find 'fa-NAME' ending with " name = re.findall(r'fa-[^""]*', line)[0] # Expects to find '\fSYMBOL' ending with " symbol = re.findall(r'\\f[^"]*', line)[0][1:].upper() camel_case = [w.capitalize() for w in name.split('-')] camel_case[0] = camel_case[0].lower() camel_name = ''.join(camel_case) name = name.lstrip('fa-') print('\expandafter\def\csname faicon@{name}\endcsname ' '{{\symbol{{"{symbol}}}}} \def\{camel_name} ' '{{{{\FA\csname faicon@{name}\endcsname}}}}'.format(name=name, camel_name=camel_name, symbol=symbol))
import re # Input file created from http://astronautweb.co/snippet/font-awesome/ INPUT_FILE = 'htmlfontawesome.txt' OUTPUT_FILE = 'fontawesome.sty' with open(INPUT_FILE) as r, open(OUTPUT_FILE, 'w') as w: for line in r: # Expects to find 'fa-NAME' ending with " name = re.findall(r'fa-[^""]*', line)[0] # Expects to find '\fSYMBOL' ending with " symbol = re.findall(r'\\f[^"]*', line)[0][1:].upper() camel_case = [w.capitalize() for w in name.split('-')] camel_case[0] = camel_case[0].lower() camel_name = ''.join(camel_case) name = name.lstrip('fa-') print('\expandafter\def\csname faicon@{name}\endcsname ' '{{\symbol{{"{symbol}}}}} \def\{camel_name} ' '{{{{\FA\csname faicon@{name}\endcsname}}}}'.format(name=name, camel_name=camel_name, symbol=symbol), file=w)
Write output to .sty file.
Write output to .sty file.
Python
mit
posquit0/latex-fontawesome
--- +++ @@ -2,8 +2,9 @@ # Input file created from http://astronautweb.co/snippet/font-awesome/ INPUT_FILE = 'htmlfontawesome.txt' +OUTPUT_FILE = 'fontawesome.sty' -with open(INPUT_FILE) as r: +with open(INPUT_FILE) as r, open(OUTPUT_FILE, 'w') as w: for line in r: # Expects to find 'fa-NAME' ending with " name = re.findall(r'fa-[^""]*', line)[0] @@ -18,4 +19,4 @@ print('\expandafter\def\csname faicon@{name}\endcsname ' '{{\symbol{{"{symbol}}}}} \def\{camel_name} ' '{{{{\FA\csname faicon@{name}\endcsname}}}}'.format(name=name, - camel_name=camel_name, symbol=symbol)) + camel_name=camel_name, symbol=symbol), file=w)
d1e2aacb7926a7e751cd27eb562b2c5d86f7e1e8
opal/tests/test_core_test_runner.py
opal/tests/test_core_test_runner.py
""" Unittests fror opal.core.test_runner """ import ffs from mock import MagicMock, patch from opal.core.test import OpalTestCase from opal.core import test_runner class RunPyTestsTestCase(OpalTestCase): @patch('subprocess.check_call') def test_run_tests(self, check_call): mock_args = MagicMock(name="args") mock_args.userland_here = ffs.Path('.') mock_args.coverage = False mock_args.test = None test_runner._run_py_tests(mock_args) check_call.assert_called_once_with(['python', 'runtests.py']) class RunJSTestsTestCase(OpalTestCase): pass class RunTestsTestCase(OpalTestCase): pass
""" Unittests fror opal.core.test_runner """ import ffs from mock import MagicMock, patch from opal.core.test import OpalTestCase from opal.core import test_runner class RunPyTestsTestCase(OpalTestCase): @patch('subprocess.check_call') def test_run_tests(self, check_call): mock_args = MagicMock(name="args") mock_args.userland_here = ffs.Path('.') mock_args.coverage = False mock_args.test = None test_runner._run_py_tests(mock_args) check_call.assert_called_once_with(['python', 'runtests.py']) @patch('subprocess.check_call') def test_run_tests_with_test_arg(self, check_call): mock_args = MagicMock(name="args") mock_args.userland_here = ffs.Path('.') mock_args.coverage = False mock_args.test = 'opal.tests.foo' test_runner._run_py_tests(mock_args) check_call.assert_called_once_with(['python', 'runtests.py', 'opal.tests.foo']) class RunJSTestsTestCase(OpalTestCase): pass class RunTestsTestCase(OpalTestCase): pass
Add test for opal test py -t
Add test for opal test py -t
Python
agpl-3.0
khchine5/opal,khchine5/opal,khchine5/opal
--- +++ @@ -19,6 +19,17 @@ test_runner._run_py_tests(mock_args) check_call.assert_called_once_with(['python', 'runtests.py']) + @patch('subprocess.check_call') + def test_run_tests_with_test_arg(self, check_call): + mock_args = MagicMock(name="args") + mock_args.userland_here = ffs.Path('.') + mock_args.coverage = False + mock_args.test = 'opal.tests.foo' + test_runner._run_py_tests(mock_args) + check_call.assert_called_once_with(['python', 'runtests.py', 'opal.tests.foo']) + + + class RunJSTestsTestCase(OpalTestCase): pass
267271f1d875e13b8a162976891bc8f3298fe8ba
stdnumfield/models.py
stdnumfield/models.py
# coding=utf-8 from django.core.exceptions import ImproperlyConfigured from django.db import models from django.utils.itercompat import is_iterable from six import string_types from . import settings from .forms import StdnumField __all__ = [ 'StdNumField', ] class StdNumField(models.CharField): """Model field that can store an stdnum value""" def __init__(self, formats, *args, **kwargs): if formats is None: raise ImproperlyConfigured('StdNumField defined without formats') if not is_iterable(formats) or isinstance(formats, string_types): formats = [formats] for format in formats: if format not in settings.DEFAULT_FORMATS: raise ValueError( 'Unknown format for StdNumField: "{}". Is it missing from ' 'settings.STDNUMFIELD["DEFAULT_FORMATS"]?'.format( format, )) self.formats = formats # TODO make dynamic when/if stdnum provides this data: kwargs["max_length"] = 254 super(StdNumField, self).__init__(*args, **kwargs) def deconstruct(self): name, path, args, kwargs = super(StdNumField, self).deconstruct() kwargs['formats'] = self.formats del kwargs["max_length"] return name, path, args, kwargs def get_internal_type(self): return 'CharField' def formfield(self, **kwargs): defaults = {'form_class': StdnumField} defaults.update(kwargs) return super(StdNumField, self).formfield(**defaults)
# coding=utf-8 from django.core.exceptions import ImproperlyConfigured from django.db import models from django.utils.itercompat import is_iterable from six import string_types from . import settings from .forms import StdnumField __all__ = [ 'StdNumField', ] class StdNumField(models.CharField): """Model field that can store an stdnum value""" def __init__(self, formats, *args, **kwargs): if formats is None: raise ImproperlyConfigured('StdNumField defined without formats') if not is_iterable(formats) or isinstance(formats, string_types): formats = [formats] for format in formats: if format not in settings.DEFAULT_FORMATS: raise ValueError( 'Unknown format for StdNumField: "{}". Is it missing from ' 'settings.STDNUMFIELD["DEFAULT_FORMATS"]?'.format( format, )) self.formats = formats # TODO make dynamic when/if stdnum provides this data: kwargs["max_length"] = 254 super(StdNumField, self).__init__(*args, **kwargs) def deconstruct(self): name, path, args, kwargs = super(StdNumField, self).deconstruct() kwargs['formats'] = self.formats del kwargs["max_length"] return name, path, args, kwargs def get_internal_type(self): return 'CharField' def formfield(self, **kwargs): defaults = { 'form_class': StdnumField, 'formats': self.formats, } defaults.update(kwargs) return super(StdNumField, self).formfield(**defaults)
Fix form field custom kwarg
Fix form field custom kwarg
Python
unlicense
frnhr/django-stdnumfield,frnhr/django-stdnumfield,frnhr/django-stdnumfield
--- +++ @@ -43,6 +43,9 @@ return 'CharField' def formfield(self, **kwargs): - defaults = {'form_class': StdnumField} + defaults = { + 'form_class': StdnumField, + 'formats': self.formats, + } defaults.update(kwargs) return super(StdNumField, self).formfield(**defaults)
9637218c8b544c397bcd5d433de47cafbfad973d
octodns/source/base.py
octodns/source/base.py
# # # from __future__ import absolute_import, division, print_function, \ unicode_literals class BaseSource(object): def __init__(self, id): self.id = id if not getattr(self, 'log', False): raise NotImplementedError('Abstract base class, log property ' 'missing') if not hasattr(self, 'SUPPORTS_GEO'): raise NotImplementedError('Abstract base class, SUPPORTS_GEO ' 'property missing') if not hasattr(self, 'SUPPORTS'): raise NotImplementedError('Abstract base class, SUPPORTS ' 'property missing') def populate(self, zone, target=False): ''' Loads all zones the provider knows about When `target` is True the populate call is being made to load the current state of the provider. When `lenient` is True the populate call may skip record validation and do a "best effort" load of data. That will allow through some common, but not best practices stuff that we otherwise would reject. E.g. no trailing . or mising escapes for ;. ''' raise NotImplementedError('Abstract base class, populate method ' 'missing') def supports(self, record): return record._type in self.SUPPORTS def __repr__(self): return self.__class__.__name__
# # # from __future__ import absolute_import, division, print_function, \ unicode_literals class BaseSource(object): def __init__(self, id): self.id = id if not getattr(self, 'log', False): raise NotImplementedError('Abstract base class, log property ' 'missing') if not hasattr(self, 'SUPPORTS_GEO'): raise NotImplementedError('Abstract base class, SUPPORTS_GEO ' 'property missing') if not hasattr(self, 'SUPPORTS'): raise NotImplementedError('Abstract base class, SUPPORTS ' 'property missing') def populate(self, zone, target=False, lenient=False): ''' Loads all zones the provider knows about When `target` is True the populate call is being made to load the current state of the provider. When `lenient` is True the populate call may skip record validation and do a "best effort" load of data. That will allow through some common, but not best practices stuff that we otherwise would reject. E.g. no trailing . or mising escapes for ;. ''' raise NotImplementedError('Abstract base class, populate method ' 'missing') def supports(self, record): return record._type in self.SUPPORTS def __repr__(self): return self.__class__.__name__
Add lenient to abstract BaseSource signature
Add lenient to abstract BaseSource signature
Python
mit
vanbroup/octodns,vanbroup/octodns,h-hwang/octodns,h-hwang/octodns
--- +++ @@ -20,7 +20,7 @@ raise NotImplementedError('Abstract base class, SUPPORTS ' 'property missing') - def populate(self, zone, target=False): + def populate(self, zone, target=False, lenient=False): ''' Loads all zones the provider knows about
487acc33f8086a889193e5995424e7bfdbb208ce
django_project/frontend/tests/test_views.py
django_project/frontend/tests/test_views.py
# -*- coding: utf-8 -*- from django.test import TestCase, Client from django.core.urlresolvers import reverse class TestViews(TestCase): def setUp(self): self.client = Client() def test_home_view(self): resp = self.client.get(reverse('home')) self.assertEqual(resp.status_code, 200) self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'main.html', u'base.html', u'pipeline/css.html', u'pipeline/js.html', u'pipeline/js.html' ] )
# -*- coding: utf-8 -*- from django.test import TestCase, Client from django.core.urlresolvers import reverse from django.conf import settings class TestViews(TestCase): def setUp(self): self.client = Client() def test_home_view(self): resp = self.client.get(reverse('home')) self.assertEqual(resp.status_code, 200) self.assertEqual(resp.context['debug'], False) self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'main.html', u'base.html', u'pipeline/css.html', u'pipeline/js.html', u'pipeline/js.html' ] ) def test_home_view_no_googleanalytics(self): # specifically set DEBUG to True settings.DEBUG = True resp = self.client.get(reverse('home')) self.assertEqual(resp.status_code, 200) self.assertEqual(resp.context['debug'], True) self.assertTrue(resp.content.find('GoogleAnalyticsObject') == -1) self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'main.html', u'base.html', u'pipeline/css.html', u'pipeline/js.html', u'pipeline/js.html' ] ) def test_about_view(self): resp = self.client.get(reverse('about')) self.assertEqual(resp.status_code, 200) self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'about.html' ] ) def test_help_view(self): resp = self.client.get(reverse('help')) self.assertEqual(resp.status_code, 200) self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'help.html' ] )
Add missing tests for frontend views
Add missing tests for frontend views
Python
bsd-2-clause
ismailsunni/healthsites,ismailsunni/healthsites,ismailsunni/healthsites,ismailsunni/healthsites
--- +++ @@ -1,6 +1,7 @@ # -*- coding: utf-8 -*- from django.test import TestCase, Client from django.core.urlresolvers import reverse +from django.conf import settings class TestViews(TestCase): @@ -11,9 +12,49 @@ resp = self.client.get(reverse('home')) self.assertEqual(resp.status_code, 200) + + self.assertEqual(resp.context['debug'], False) + self.assertListEqual( [tmpl.name for tmpl in resp.templates], [ 'main.html', u'base.html', u'pipeline/css.html', u'pipeline/js.html', u'pipeline/js.html' ] ) + + def test_home_view_no_googleanalytics(self): + # specifically set DEBUG to True + settings.DEBUG = True + + resp = self.client.get(reverse('home')) + + self.assertEqual(resp.status_code, 200) + + self.assertEqual(resp.context['debug'], True) + self.assertTrue(resp.content.find('GoogleAnalyticsObject') == -1) + self.assertListEqual( + [tmpl.name for tmpl in resp.templates], [ + 'main.html', u'base.html', u'pipeline/css.html', + u'pipeline/js.html', u'pipeline/js.html' + ] + ) + + def test_about_view(self): + resp = self.client.get(reverse('about')) + + self.assertEqual(resp.status_code, 200) + self.assertListEqual( + [tmpl.name for tmpl in resp.templates], [ + 'about.html' + ] + ) + + def test_help_view(self): + resp = self.client.get(reverse('help')) + + self.assertEqual(resp.status_code, 200) + self.assertListEqual( + [tmpl.name for tmpl in resp.templates], [ + 'help.html' + ] + )
db01eb72829db32c87a167c9b3529577a028ec54
example_project/example_project/settings.py
example_project/example_project/settings.py
# Django settings for example_project project. DEBUG = True TEMPLATE_DEBUG = DEBUG DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'example_project.sqlite3', } } SITE_ID = 1 SECRET_KEY = 'u%38dln@$1!7w#cxi4np504^sa3_skv5aekad)jy_u0v2mc+nr' TEMPLATE_LOADERS = ( 'django.template.loaders.filesystem.Loader', 'django.template.loaders.app_directories.Loader', ) ROOT_URLCONF = 'example_project.urls' STATIC_URL = '/static/' DJANGO_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.admin', ) THIRD_PARTY_APPS = ( 'embed_video', ) LOCAL_APPS = ( 'example_project', 'posts', ) INSTALLED_APPS = DJANGO_APPS + THIRD_PARTY_APPS + LOCAL_APPS
# Django settings for example_project project. DEBUG = True TEMPLATE_DEBUG = DEBUG DATABASES = { 'default': { 'ENGINE': 'django.db.backends.sqlite3', 'NAME': 'example_project.sqlite3', } } SITE_ID = 1 SECRET_KEY = 'u%38dln@$1!7w#cxi4np504^sa3_skv5aekad)jy_u0v2mc+nr' TEMPLATE_LOADERS = ( 'django.template.loaders.filesystem.Loader', 'django.template.loaders.app_directories.Loader', ) APPEND_SLASH = True ROOT_URLCONF = 'example_project.urls' STATIC_URL = '/static/' DJANGO_APPS = ( 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.admin', ) THIRD_PARTY_APPS = ( 'embed_video', ) LOCAL_APPS = ( 'example_project', 'posts', ) INSTALLED_APPS = DJANGO_APPS + THIRD_PARTY_APPS + LOCAL_APPS
Append slash to urls in example project
Append slash to urls in example project
Python
mit
yetty/django-embed-video,mpachas/django-embed-video,yetty/django-embed-video,jazzband/django-embed-video,jazzband/django-embed-video,mpachas/django-embed-video
--- +++ @@ -19,6 +19,7 @@ 'django.template.loaders.app_directories.Loader', ) +APPEND_SLASH = True ROOT_URLCONF = 'example_project.urls'
1c6392889a5393b9681d18aab5294c1f1927730a
__init__.py
__init__.py
_VERSION = 'CVS' _TEMP_DIR = '.SloppyCell' import logging logging.basicConfig() logger = logging.getLogger('__init__') # Check for debugging option. I tried using optparse for this, but ran into # issues with ipython and mpirun, both of which pollute sys.argv. import sys for arg in sys.argv: if arg.startswith('--debugSC'): words = arg.split('=') import Utility if len(words) == 2: Utility.enable_debugging_msgs(words[1]) else: Utility.enable_debugging_msgs(None) import os currdir = os.getcwd() try: import pypar os.chdir(currdir) HAVE_PYPAR = True num_procs = pypar.size() my_rank = pypar.rank() my_host = pypar.get_processor_name() import atexit atexit.register(pypar.finalize) except ImportError: os.chdir(currdir) HAVE_PYPAR = False num_procs = 1 my_rank = 0 import socket my_host = socket.gethostname() logger.debug('Node %i is on host %s.' % (my_rank, my_host)) import os if my_rank == 0 and not os.path.isdir(_TEMP_DIR): os.mkdir(_TEMP_DIR) import OldScipySupport
_VERSION = 'CVS' import os _TEMP_DIR = os.path.join(os.getcwd(), '.SloppyCell') import logging logging.basicConfig() logger = logging.getLogger('__init__') # Check for debugging option. I tried using optparse for this, but ran into # issues with ipython and mpirun, both of which pollute sys.argv. import sys for arg in sys.argv: if arg.startswith('--debugSC'): words = arg.split('=') import Utility if len(words) == 2: Utility.enable_debugging_msgs(words[1]) else: Utility.enable_debugging_msgs(None) currdir = os.getcwd() try: import pypar os.chdir(currdir) HAVE_PYPAR = True num_procs = pypar.size() my_rank = pypar.rank() my_host = pypar.get_processor_name() import atexit atexit.register(pypar.finalize) except ImportError: os.chdir(currdir) HAVE_PYPAR = False num_procs = 1 my_rank = 0 import socket my_host = socket.gethostname() logger.debug('Node %i is on host %s.' % (my_rank, my_host)) if my_rank == 0 and not os.path.isdir(_TEMP_DIR): os.mkdir(_TEMP_DIR) import OldScipySupport
Handle temp directory with absolute path
Handle temp directory with absolute path
Python
bsd-3-clause
GutenkunstLab/SloppyCell,GutenkunstLab/SloppyCell
--- +++ @@ -1,5 +1,6 @@ _VERSION = 'CVS' -_TEMP_DIR = '.SloppyCell' +import os +_TEMP_DIR = os.path.join(os.getcwd(), '.SloppyCell') import logging logging.basicConfig() @@ -16,7 +17,6 @@ Utility.enable_debugging_msgs(words[1]) else: Utility.enable_debugging_msgs(None) -import os currdir = os.getcwd() try: @@ -37,7 +37,6 @@ my_host = socket.gethostname() logger.debug('Node %i is on host %s.' % (my_rank, my_host)) -import os if my_rank == 0 and not os.path.isdir(_TEMP_DIR): os.mkdir(_TEMP_DIR)
d38fee106c00ed20b9a1ed3f38c057393576f6ea
tmaps/defaultconfig.py
tmaps/defaultconfig.py
import logging import datetime DEBUG = True # Override this key with a secret one SECRET_KEY = 'default_secret_key' HASHIDS_SALT = 'default_secret_salt' ## Authentication JWT_EXPIRATION_DELTA = datetime.timedelta(days=2) JWT_NOT_BEFORE_DELTA = datetime.timedelta(seconds=0) ## Database SQLALCHEMY_DATABASE_URI = None SQLALCHEMY_TRACK_MODIFICATIONS = True ## Logging LOG_FILE = 'tissuemaps.log' LOG_LEVEL = logging.INFO LOG_MAX_BYTES = 2048000 # 2048KB LOG_N_BACKUPS = 10 ## Other # This should be set to true in the production config when using NGINX USE_X_SENDFILE = False REDIS_URL = 'redis://localhost:6379'
import logging import datetime DEBUG = False # Override this key with a secret one SECRET_KEY = 'default_secret_key' HASHIDS_SALT = 'default_secret_salt' ## Authentication JWT_EXPIRATION_DELTA = datetime.timedelta(days=2) JWT_NOT_BEFORE_DELTA = datetime.timedelta(seconds=0) ## Database SQLALCHEMY_DATABASE_URI = None SQLALCHEMY_TRACK_MODIFICATIONS = True ## Logging LOG_FILE = 'tissuemaps.log' LOG_LEVEL = logging.INFO LOG_MAX_BYTES = 2048000 # 2048KB LOG_N_BACKUPS = 10 ## Other # This should be set to true in the production config when using NGINX USE_X_SENDFILE = False REDIS_URL = 'redis://localhost:6379'
Remove spark settings from default config file
Remove spark settings from default config file
Python
agpl-3.0
TissueMAPS/TmServer
--- +++ @@ -1,7 +1,7 @@ import logging import datetime -DEBUG = True +DEBUG = False # Override this key with a secret one SECRET_KEY = 'default_secret_key'
71d56354fb053c6cef3dc2c8960f78f588327114
project/views.py
project/views.py
#! coding: utf-8 from django.shortcuts import render_to_response, render from django.http import HttpResponseRedirect from django.contrib.auth import login from forms import LoginForm def index(request): return render_to_response('index.html', {}) def login_view(request): if request.method == 'POST': form = LoginForm(request.POST) if form.is_valid(): user = form.cleaned_data['user'] if user is not None and user.is_active: login(request, user) return HttpResponseRedirect('/') else: form = LoginForm() return render(request, 'login.html', { 'form': form, })
#! coding: utf-8 from django.shortcuts import render_to_response, render from django.http import HttpResponseRedirect from django.contrib.auth import login from forms import LoginForm def index(request): return render_to_response('index.html', {}) def login_view(request): if request.method == 'POST': form = LoginForm(request.POST) if form.is_valid(): user = form.cleaned_data['user'] if user is not None and user.is_active: request.session['password'] = form.cleaned_data['password'] login(request, user) return HttpResponseRedirect('/') else: form = LoginForm() return render(request, 'login.html', { 'form': form, })
Store password in session after successful login.
Store password in session after successful login.
Python
agpl-3.0
InScience/DAMIS-old,InScience/DAMIS-old
--- +++ @@ -15,6 +15,7 @@ if form.is_valid(): user = form.cleaned_data['user'] if user is not None and user.is_active: + request.session['password'] = form.cleaned_data['password'] login(request, user) return HttpResponseRedirect('/') else:
9b55af4eb1d40517fce310b0751713dc8448f13f
main.py
main.py
import os import logging import gevent from flask import Flask, render_template, url_for, redirect from flask_sockets import Sockets import io app = Flask(__name__) path = os.getcwd() app.config['DEBUG'] = True sockets = Sockets(app) @app.route('/', methods=['GET', 'POST']) def main(): return redirect(url_for('static', filename='index.html')) @sockets.route('/submit') def submit(ws): while not ws.closed: gevent.sleep() data = ws.receive() if data: ws.send("ayy lmao") wave_file = io.open("test.wav", "wb") wave_file.write(data) wave_file.close()
import os import logging import gevent from flask import Flask, render_template, url_for, redirect from flask_sockets import Sockets import io import string import random app = Flask(__name__) path = os.getcwd() app.config['DEBUG'] = True sockets = Sockets(app) def rand_id(size=8): return ''.join(random.SystemRandom().choice(string.ascii_uppercase + string.digits) for _ in range(size)) @app.route('/', methods=['GET', 'POST']) def main(): return redirect(url_for('static', filename='index.html')) @sockets.route('/submit') def submit(ws): user_id = rand_id() while not ws.closed: gevent.sleep() data = ws.receive() if data and data != "start": file_name = "_".join(["file", user_id, rand_id()]) wave_file = io.open(file_name, "wb") wave_file.write(data) wave_file.close() # process_file(file_name)
Add unique id for each wav file
Add unique id for each wav file
Python
mit
j-salazar/mchacks15,j-salazar/mchacks15,j-salazar/mchacks15
--- +++ @@ -4,11 +4,16 @@ from flask import Flask, render_template, url_for, redirect from flask_sockets import Sockets import io +import string +import random app = Flask(__name__) path = os.getcwd() app.config['DEBUG'] = True sockets = Sockets(app) + +def rand_id(size=8): + return ''.join(random.SystemRandom().choice(string.ascii_uppercase + string.digits) for _ in range(size)) @app.route('/', methods=['GET', 'POST']) def main(): @@ -16,11 +21,13 @@ @sockets.route('/submit') def submit(ws): + user_id = rand_id() while not ws.closed: gevent.sleep() data = ws.receive() - if data: - ws.send("ayy lmao") - wave_file = io.open("test.wav", "wb") + if data and data != "start": + file_name = "_".join(["file", user_id, rand_id()]) + wave_file = io.open(file_name, "wb") wave_file.write(data) wave_file.close() + # process_file(file_name)
39cfea1b0528822720b88b890ae84fdf120826ff
bibliopixel/util/threads/compose_events.py
bibliopixel/util/threads/compose_events.py
import functools, threading def compose_events(events, condition=all): """Compose a sequence of events into one event. Arguments: events: a sequence of objects looking like threading.Event condition: a function taking a sequence of bools and returning a bool. """ events = list(events) master_event = threading.Event() def changed(): if condition(e.is_set() for e in events): master_event.set() else: master_event.clear() def add_changed(f): @functools.wraps(f) def wrapped(): f() changed() return wrapped for e in events: e.set = add_changed(e.set) e.clear = add_changed(e.clear) changed() return master_event
import functools, threading def compose_events(events, condition=all): """Compose a sequence of events into one event. Arguments: events: a sequence of objects looking like threading.Event condition: a function taking a sequence of bools and returning a bool. """ events = list(events) master_event = threading.Event() def changed(): if condition(e.is_set() for e in events): master_event.set() else: master_event.clear() def add_changed(f): @functools.wraps(f) def wrapped(): f() changed() return wrapped for e in events: e.set = add_changed(e.set) e.clear = add_changed(e.clear) changed() return master_event
Add an empty line for style
Add an empty line for style
Python
mit
rec/BiblioPixel,ManiacalLabs/BiblioPixel,rec/BiblioPixel,rec/BiblioPixel,rec/BiblioPixel,ManiacalLabs/BiblioPixel,ManiacalLabs/BiblioPixel,ManiacalLabs/BiblioPixel
--- +++ @@ -21,6 +21,7 @@ def wrapped(): f() changed() + return wrapped for e in events:
c65a475c38a611cbf55f2dacbe22ccd50597c9ed
tests/test_database/test_sql/test_median.py
tests/test_database/test_sql/test_median.py
import unittest from tkp.db import execute, rollback class testMedian(unittest.TestCase): def setUp(self): try: execute('drop table median_test') except: rollback() execute('create table median_test (i int, f float)') execute('insert into median_test values (1, 1.1)') execute('insert into median_test values (2, 2.1)') execute('insert into median_test values (3, 3.1)') def tearDown(self): rollback() def test_median(self): cursor = execute('select median(i), median(f) from median_test') median_i, median_f = cursor.fetchall()[0] self.assertEqual(median_i, 2) self.assertEqual(median_f, 2.1)
import unittest import tkp from tkp.db import execute, rollback, Database from tkp.testutil import db_subs from numpy import median class testMedian(unittest.TestCase): def setUp(self): self.database = tkp.db.Database() self.dataset = tkp.db.DataSet(database=self.database, data={'description':"Median test" + self._testMethodName}) self.n_images = 5 self.im_params = db_subs.generate_timespaced_dbimages_data(self.n_images) for idx, impar in enumerate(self.im_params): impar['rms_max'] = (idx+1)*1e-4 self.image_ids = [] for img_pars in self.im_params: image,_,_ = db_subs.insert_image_and_simulated_sources( self.dataset,img_pars,[], new_source_sigma_margin=3) self.image_ids.append(image.id) def test_median(self): if Database().engine == 'monetdb': qry = (""" SELECT sys.median(id) as median_id ,sys.median(rms_max) as median_rms_max FROM image WHERE dataset = %(dataset_id)s """) else: qry = (""" SELECT median(id) as median_id ,median(rms_max) as median_rms_max FROM image WHERE dataset = %(dataset_id)s """) cursor = execute(qry, {'dataset_id': self.dataset.id}) results = db_subs.get_db_rows_as_dicts(cursor) # self.assertAlmostEqual(results[0]['median_id'], median(self.image_ids)) self.assertAlmostEqual(results[0]['median_rms_max'], median([p['rms_max'] for p in self.im_params]))
Use MonetDB friendly median query syntax in unit test.
Use MonetDB friendly median query syntax in unit test.
Python
bsd-2-clause
transientskp/tkp,mkuiack/tkp,bartscheers/tkp,mkuiack/tkp,transientskp/tkp,bartscheers/tkp
--- +++ @@ -1,24 +1,48 @@ import unittest -from tkp.db import execute, rollback - +import tkp +from tkp.db import execute, rollback, Database +from tkp.testutil import db_subs +from numpy import median class testMedian(unittest.TestCase): def setUp(self): - try: - execute('drop table median_test') - except: - rollback() + self.database = tkp.db.Database() - execute('create table median_test (i int, f float)') - execute('insert into median_test values (1, 1.1)') - execute('insert into median_test values (2, 2.1)') - execute('insert into median_test values (3, 3.1)') + self.dataset = tkp.db.DataSet(database=self.database, + data={'description':"Median test" + + self._testMethodName}) + self.n_images = 5 - def tearDown(self): - rollback() + self.im_params = db_subs.generate_timespaced_dbimages_data(self.n_images) + for idx, impar in enumerate(self.im_params): + impar['rms_max'] = (idx+1)*1e-4 + + self.image_ids = [] + for img_pars in self.im_params: + image,_,_ = db_subs.insert_image_and_simulated_sources( + self.dataset,img_pars,[], + new_source_sigma_margin=3) + self.image_ids.append(image.id) + def test_median(self): - cursor = execute('select median(i), median(f) from median_test') - median_i, median_f = cursor.fetchall()[0] - self.assertEqual(median_i, 2) - self.assertEqual(median_f, 2.1) + if Database().engine == 'monetdb': + qry = (""" + SELECT sys.median(id) as median_id + ,sys.median(rms_max) as median_rms_max + FROM image + WHERE dataset = %(dataset_id)s + """) + else: + qry = (""" + SELECT median(id) as median_id + ,median(rms_max) as median_rms_max + FROM image + WHERE dataset = %(dataset_id)s + """) + cursor = execute(qry, {'dataset_id': self.dataset.id}) + results = db_subs.get_db_rows_as_dicts(cursor) + # self.assertAlmostEqual(results[0]['median_id'], median(self.image_ids)) + self.assertAlmostEqual(results[0]['median_rms_max'], + median([p['rms_max'] for p in self.im_params])) +
0f5f5677ac2a1aa10067cbb509de28752fa106c0
response.py
response.py
from __future__ import division import numpy as np import matplotlib.pyplot as plt def deconv(x, y, fs): X = np.fft.fft(x) Y = np.fft.fft(y) H = Y / X h = np.fft.ifft(H) print("h =", h) # complex vector? t = np.arange(len(x)) / fs plt.plot(t, h.real) plt.grid() plt.title("impulse response") plt.show()
"""Response calculation.""" from __future__ import division import numpy as np def calculate(signal_excitation, signal_out): """Function returns impulse response.""" X = np.fft.fft(signal_excitation) Y = np.fft.fft(signal_out) H = Y / X h = np.fft.ifft(H) return h
Change function content and pep257
Change function content and pep257
Python
mit
franzpl/sweep,spatialaudio/sweep
--- +++ @@ -1,16 +1,12 @@ +"""Response calculation.""" from __future__ import division import numpy as np -import matplotlib.pyplot as plt -def deconv(x, y, fs): - X = np.fft.fft(x) - Y = np.fft.fft(y) +def calculate(signal_excitation, signal_out): + """Function returns impulse response.""" + X = np.fft.fft(signal_excitation) + Y = np.fft.fft(signal_out) H = Y / X h = np.fft.ifft(H) - print("h =", h) # complex vector? - t = np.arange(len(x)) / fs - plt.plot(t, h.real) - plt.grid() - plt.title("impulse response") - plt.show() + return h
0ede19a4f2c9c6f01db0040d9d108eb0a0b2558c
py/kafka-tmdb.py
py/kafka-tmdb.py
import json from get_tmdb import GetTMDB from kafka import KafkaConsumer try: from GLOBALS import KAFKA_BROKER, TMDB_API except ImportError: print('Get it somewhere else') class CollectTMDB(object): def __init__(self, ): self.tmdb = GetTMDB(TMDB_API) self.consumer = KafkaConsumer(group_id='tmdb', bootstrap_servers=['{}:9092'.format(KAFKA_BROKER)]) self.consumer.subscribe(pattern='tmdb') def run(self): ''' Collects a message from the topic 'tmdb'. This message is a json containing all the information collected from the apis further up the stream. We get the imdb_id from this data and pass it to the tmdb api. We append the information from the tmdb_api to the msg we collected, and then pass it to the next topic. ''' for message in self.consumer: # message value and key are raw bytes -- decode if necessary! # e.g., for unicode: `message.value.decode('utf-8') msg_data = json.loads(message.value) imdb_id = msg_data['imdb_id'] tmdb_data = GetTMDB.get_info(imdb_id) msg_data.extend(tmdb_data)
import json from get_tmdb import GetTMDB from kafka import KafkaConsumer, KafkaProducer try: from GLOBALS import KAFKA_BROKER, TMDB_API except ImportError: print('Get it somewhere else') class CollectTMDB(object): def __init__(self, ): self.tmdb = GetTMDB(TMDB_API) self.producer = KafkaProducer(bootstrap_servers=KAFKA_BROKER) self.consumer = KafkaConsumer(group_id='tmdb', bootstrap_servers=KAFKA_BROKER) self.consumer.subscribe(pattern='omdb') def run(self): ''' Collects a message from the topic 'tmdb'. This message is a json containing all the information collected from the apis further up the stream. We get the imdb_id from this data and pass it to the tmdb api. We append the information from the tmdb_api to the msg we collected, and then pass it to the next topic. ''' for message in self.consumer: # message value and key are raw bytes -- decode if necessary! # e.g., for unicode: `message.value.decode('utf-8') msg_data = json.loads(message.value) imdb_id = msg_data['imdb_id'] tmdb_data = GetTMDB.get_info(imdb_id) msg_data.extend(tmdb_data) self.producer.send('tmdb', json.dumps(msg_data))
Change KAFKA_BROKER parameter, added a send producer
Change KAFKA_BROKER parameter, added a send producer
Python
mit
kinoreel/kino-gather
--- +++ @@ -1,7 +1,7 @@ import json from get_tmdb import GetTMDB -from kafka import KafkaConsumer +from kafka import KafkaConsumer, KafkaProducer try: from GLOBALS import KAFKA_BROKER, TMDB_API @@ -12,9 +12,10 @@ def __init__(self, ): self.tmdb = GetTMDB(TMDB_API) + self.producer = KafkaProducer(bootstrap_servers=KAFKA_BROKER) self.consumer = KafkaConsumer(group_id='tmdb', - bootstrap_servers=['{}:9092'.format(KAFKA_BROKER)]) - self.consumer.subscribe(pattern='tmdb') + bootstrap_servers=KAFKA_BROKER) + self.consumer.subscribe(pattern='omdb') def run(self): ''' @@ -30,3 +31,4 @@ imdb_id = msg_data['imdb_id'] tmdb_data = GetTMDB.get_info(imdb_id) msg_data.extend(tmdb_data) + self.producer.send('tmdb', json.dumps(msg_data))
072d8fd3ccff957b427fca5e61b5a410a6762615
pulldb/publishers.py
pulldb/publishers.py
# Copyright 2013 Russell Heilling from google.appengine.ext import ndb class Publisher(ndb.Model): '''Publisher object in datastore. Holds publisher data. ''' identifier = ndb.IntegerProperty() name = ndb.StringProperty() image = ndb.StringProperty() def fetch_or_store(identifier, publisher): publisher_key = Publisher.query(Publisher.identifier==identifier).get() if not publisher_key: publisher_key = Publisher(identifier=publisher.id, name=publisher.name, image=publisher.image['tiny_url']) publisher_key.put() return publisher_key
# Copyright 2013 Russell Heilling from google.appengine.ext import ndb class Publisher(ndb.Model): '''Publisher object in datastore. Holds publisher data. ''' identifier = ndb.IntegerProperty() name = ndb.StringProperty() image = ndb.StringProperty() def fetch_or_store(identifier, publisher): publisher_key = Publisher.query(Publisher.identifier==identifier).get() if not publisher_key: publisher_key = Publisher(identifier=publisher.id, name=publisher.name) if publisher.image: publisher_key.image=publisher.image.get('tiny_url') publisher_key.put() return publisher_key
Handle null image attribute on publisher
Handle null image attribute on publisher
Python
mit
xchewtoyx/pulldb
--- +++ @@ -14,7 +14,8 @@ def fetch_or_store(identifier, publisher): publisher_key = Publisher.query(Publisher.identifier==identifier).get() if not publisher_key: - publisher_key = Publisher(identifier=publisher.id, name=publisher.name, - image=publisher.image['tiny_url']) + publisher_key = Publisher(identifier=publisher.id, name=publisher.name) + if publisher.image: + publisher_key.image=publisher.image.get('tiny_url') publisher_key.put() return publisher_key
4973656e6e569808fc9c7b50f52e67aae2c7b547
billjobs/tests/tests_export_account_email.py
billjobs/tests/tests_export_account_email.py
from django.test import TestCase from django.contrib.admin.sites import AdminSite from billjobs.admin import UserAdmin class EmailExportTestCase(TestCase): """ Tests for email account export """ def test_method_is_avaible(self): """ Test admin can select the action in dropdown list """ self.assertTrue(hasattr(UserAdmin, 'export_email')) def test_method_is_model_admin_action(self): """ Test method is an custom action for user admin """ self.assertTrue('export_email' in UserAdmin.actions) def test_action_has_a_short_description(self): """ Test method has a short description """ self.assertEqual(UserAdmin.export_email.short_description, 'Export email of selected users')
from django.test import TestCase from django.http import HttpResponse from django.contrib.admin.sites import AdminSite from django.contrib.auth.models import User from billjobs.admin import UserAdmin class EmailExportTestCase(TestCase): """ Tests for email account export """ def test_method_is_avaible(self): """ Test admin can select the action in dropdown list """ self.assertTrue(hasattr(UserAdmin, 'export_email')) def test_method_is_model_admin_action(self): """ Test method is an custom action for user admin """ self.assertTrue('export_email' in UserAdmin.actions) def test_action_has_a_short_description(self): """ Test method has a short description """ self.assertEqual(UserAdmin.export_email.short_description, 'Export email of selected users') def test_action_return_http_response(self): class MockRequest(object): pass site = AdminSite() user_admin = UserAdmin(User, site) query_set = User.objects.all() response = user_admin.export_email(request=MockRequest(), queryset=query_set) self.assertIsInstance(response, HttpResponse)
Test export email return an HttpResponse
Test export email return an HttpResponse
Python
mit
ioO/billjobs
--- +++ @@ -1,5 +1,7 @@ from django.test import TestCase +from django.http import HttpResponse from django.contrib.admin.sites import AdminSite +from django.contrib.auth.models import User from billjobs.admin import UserAdmin class EmailExportTestCase(TestCase): @@ -17,3 +19,12 @@ """ Test method has a short description """ self.assertEqual(UserAdmin.export_email.short_description, 'Export email of selected users') + + def test_action_return_http_response(self): + class MockRequest(object): + pass + site = AdminSite() + user_admin = UserAdmin(User, site) + query_set = User.objects.all() + response = user_admin.export_email(request=MockRequest(), queryset=query_set) + self.assertIsInstance(response, HttpResponse)
a91a2d3468cb3bfc7fdc686327770365321ef102
qa_app/challenges.py
qa_app/challenges.py
# Copyright 2016 Mirantis, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from flask import Blueprint, render_template from flask_login import login_required challenges = Blueprint('challenges', __name__) @challenges.route('/challenges', methods=['GET']) @login_required def challenges_view(): return render_template('challenges.html', page="Challenges")
# Copyright 2016 Mirantis, Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import requests from flask import Blueprint, render_template, jsonify from flask_login import login_required challenges = Blueprint('challenges', __name__) @challenges.route('/challenges', methods=['GET']) @login_required def challenges_view(): return render_template('challenges.html', page="Challenges") @challenges.route('/exercises', methods=['GET']) @login_required def api_exercises(): exercises = requests.get("http://localhost:8000/").json() result = { "exercises": [] } for current in exercises['exercises']: result['exercises'].append({ "name": current.get('name', 'unknown'), "category": current.get('answers')[0].split(".")[1], "solved": 0, "cost": 100 }) return jsonify(result)
Implement demo 'exercises' api method.
Implement demo 'exercises' api method.
Python
apache-2.0
molecul/qa_app_flask,molecul/qa_app_flask,molecul/qa_app_flask
--- +++ @@ -11,7 +11,9 @@ # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. -from flask import Blueprint, render_template +import requests + +from flask import Blueprint, render_template, jsonify from flask_login import login_required challenges = Blueprint('challenges', __name__) @@ -21,3 +23,21 @@ @login_required def challenges_view(): return render_template('challenges.html', page="Challenges") + + +@challenges.route('/exercises', methods=['GET']) +@login_required +def api_exercises(): + exercises = requests.get("http://localhost:8000/").json() + result = { + "exercises": + [] + } + for current in exercises['exercises']: + result['exercises'].append({ + "name": current.get('name', 'unknown'), + "category": current.get('answers')[0].split(".")[1], + "solved": 0, + "cost": 100 + }) + return jsonify(result)
5af61cae2ca438880357f88533cfa77ea161efac
corehq/ex-submodules/pillow_retry/admin.py
corehq/ex-submodules/pillow_retry/admin.py
from django.contrib import admin from .models import PillowError class PillowErrorAdmin(admin.ModelAdmin): model = PillowError list_display = [ 'pillow', 'doc_id', 'error_type', 'date_created', 'date_last_attempt', 'date_next_attempt' ] list_filter = ('pillow', 'error_type') admin.site.register(PillowError, PillowErrorAdmin)
from django.contrib import admin from pillow_retry.models import PillowError @admin.register(PillowError) class PillowErrorAdmin(admin.ModelAdmin): model = PillowError list_display = [ 'pillow', 'doc_id', 'error_type', 'date_created', 'date_last_attempt', 'date_next_attempt' ] list_filter = ('pillow', 'error_type') actions = [ 'delete_selected' ]
Add delete action to PillowRetry
Add delete action to PillowRetry
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
--- +++ @@ -1,7 +1,9 @@ from django.contrib import admin -from .models import PillowError + +from pillow_retry.models import PillowError +@admin.register(PillowError) class PillowErrorAdmin(admin.ModelAdmin): model = PillowError @@ -14,6 +16,6 @@ 'date_next_attempt' ] list_filter = ('pillow', 'error_type') - - -admin.site.register(PillowError, PillowErrorAdmin) + actions = [ + 'delete_selected' + ]
441e93cf96aa247a0cce36892e654de17ad44a8a
test/streamparse/cli/test_worker_uptime.py
test/streamparse/cli/test_worker_uptime.py
from __future__ import absolute_import, unicode_literals import argparse import unittest from streamparse.cli.worker_uptime import subparser_hook from nose.tools import ok_ class WorkerUptimeTestCase(unittest.TestCase): def test_subparser_hook(self): parser = argparse.ArgumentParser() subparsers = parser.add_subparsers() subparser_hook(subparsers) subcommands = parser._optionals._actions[1].choices.keys() ok_('worker-uptime' in subcommands) if __name__ == '__main__': unittest.main()
from __future__ import absolute_import, unicode_literals import argparse import unittest from streamparse.cli.worker_uptime import subparser_hook from nose.tools import ok_ class WorkerUptimeTestCase(unittest.TestCase): def test_subparser_hook(self): parser = argparse.ArgumentParser() subparsers = parser.add_subparsers() subparser_hook(subparsers) subcommands = parser._optionals._actions[1].choices.keys() ok_('worker_uptime' in subcommands) if __name__ == '__main__': unittest.main()
Fix unit test after worker-uptime changed to worker_uptime
Fix unit test after worker-uptime changed to worker_uptime
Python
apache-2.0
Parsely/streamparse,crohling/streamparse,petchat/streamparse,Parsely/streamparse,petchat/streamparse,hodgesds/streamparse,petchat/streamparse,codywilbourn/streamparse,phanib4u/streamparse,codywilbourn/streamparse,msmakhlouf/streamparse,msmakhlouf/streamparse,eric7j/streamparse,petchat/streamparse,crohling/streamparse,msmakhlouf/streamparse,msmakhlouf/streamparse,phanib4u/streamparse,petchat/streamparse,eric7j/streamparse,msmakhlouf/streamparse,hodgesds/streamparse
--- +++ @@ -16,7 +16,7 @@ subparser_hook(subparsers) subcommands = parser._optionals._actions[1].choices.keys() - ok_('worker-uptime' in subcommands) + ok_('worker_uptime' in subcommands) if __name__ == '__main__':
19d2dff39988309123ef97b4bb38a2eac6d18de1
tests/integration/api/test_sc_test_jobs.py
tests/integration/api/test_sc_test_jobs.py
from tests.base import BaseTest from tenable_io.api.models import ScTestJob class TestScTestJobsApi(BaseTest): def test_status(self, client, image): jobs = client.sc_test_jobs_api.list() assert len(jobs) > 0, u'At least one job exists.' test_job = client.sc_test_jobs_api.status(jobs[0].job_id) assert isinstance(test_job, ScTestJob), u'The method returns type.' def test_by_image(self, client, image): job = client.sc_test_jobs_api.by_image(image['id']) assert isinstance(job, ScTestJob), u'The method returns type.' def test_by_image_digest(self, client, image): job = client.sc_test_jobs_api.by_image(image['digest']) assert isinstance(job, ScTestJob), u'The method returns type.' def test_list(self, client, image): jobs = client.sc_test_jobs_api.list() assert len(jobs) > 0, u'At least one job exists.' assert isinstance(jobs[0], ScTestJob), u'The method returns job list.'
from tests.base import BaseTest from tenable_io.api.models import ScTestJob class TestScTestJobsApi(BaseTest): def test_status(self, client, image): jobs = client.sc_test_jobs_api.list() assert len(jobs) > 0, u'At least one job exists.' test_job = client.sc_test_jobs_api.status(jobs[0].job_id) assert isinstance(test_job, ScTestJob), u'The method returns type.' def test_by_image(self, client, image): job = client.sc_test_jobs_api.by_image(image['id']) assert isinstance(job, ScTestJob), u'The method returns type.' def test_by_image_digest(self, client, image): job = client.sc_test_jobs_api.by_image_digest(image['digest']) assert isinstance(job, ScTestJob), u'The method returns type.' def test_list(self, client, image): jobs = client.sc_test_jobs_api.list() assert len(jobs) > 0, u'At least one job exists.' assert isinstance(jobs[0], ScTestJob), u'The method returns job list.'
Fix for broken container security test
Fix for broken container security test
Python
mit
tenable/Tenable.io-SDK-for-Python
--- +++ @@ -16,7 +16,7 @@ assert isinstance(job, ScTestJob), u'The method returns type.' def test_by_image_digest(self, client, image): - job = client.sc_test_jobs_api.by_image(image['digest']) + job = client.sc_test_jobs_api.by_image_digest(image['digest']) assert isinstance(job, ScTestJob), u'The method returns type.' def test_list(self, client, image):
d7391bb7ef8d1cb2e900724f89f1753a7feb6fa7
rsr/cmd.py
rsr/cmd.py
import os import signal import sys from argparse import ArgumentParser import gi gi.require_version('Gtk', '3.0') gi.require_version('GtkSource', '3.0') from gi.repository import Gio, GLib from rsr import __version__ from rsr.app import Application parser = ArgumentParser(prog='runsqlrun', description='Run SQL statements') parser.add_argument( '--version', action='version', version='%(prog)s ' + __version__) # See issue3. Unfortunately this needs to be done before opening # any Oracle connection. os.environ.setdefault('NLS_LANG', '.AL32UTF8') def main(): parser.parse_args() signal.signal(signal.SIGINT, signal.SIG_DFL) GLib.set_application_name('RunSQLRun') GLib.set_prgname('runsqlrun') resource = Gio.resource_load('data/runsqlrun.gresource') Gio.Resource._register(resource) app = Application() sys.exit(app.run(sys.argv))
import os import signal import sys from argparse import ArgumentParser import gi gi.require_version('Gtk', '3.0') gi.require_version('GtkSource', '3.0') from gi.repository import Gio, GLib, Gtk from rsr import __version__ from rsr.app import Application parser = ArgumentParser(prog='runsqlrun', description='Run SQL statements') parser.add_argument( '--version', action='version', version='%(prog)s ' + __version__) # See issue3. Unfortunately this needs to be done before opening # any Oracle connection. os.environ.setdefault('NLS_LANG', '.AL32UTF8') def main(): parser.parse_args() signal.signal(signal.SIGINT, signal.SIG_DFL) GLib.set_application_name('RunSQLRun') GLib.set_prgname('runsqlrun') resource = Gio.resource_load('data/runsqlrun.gresource') Gio.Resource._register(resource) Gtk.Settings.get_default().set_property( 'gtk-application-prefer-dark-theme', True) app = Application() sys.exit(app.run(sys.argv))
Use dark theme if possible.
Use dark theme if possible.
Python
mit
andialbrecht/runsqlrun
--- +++ @@ -7,7 +7,7 @@ gi.require_version('Gtk', '3.0') gi.require_version('GtkSource', '3.0') -from gi.repository import Gio, GLib +from gi.repository import Gio, GLib, Gtk from rsr import __version__ from rsr.app import Application @@ -31,5 +31,8 @@ resource = Gio.resource_load('data/runsqlrun.gresource') Gio.Resource._register(resource) + Gtk.Settings.get_default().set_property( + 'gtk-application-prefer-dark-theme', True) + app = Application() sys.exit(app.run(sys.argv))
3b07818db48a5e3a205389051ccd9640e1079cc7
tests/lib/__init__.py
tests/lib/__init__.py
from os.path import abspath, dirname, join import sh import psycopg2 import requests PROJECT_FOLDER=dirname(dirname(abspath(__file__))) DOCKER_FOLDER=join(PROJECT_FOLDER, 'docker') def docker_compose(version, *args): sh.docker_compose('-f', 'docker/{version}/docker-compose.yml'.format(version=version), *args) def test_pg(): def callback(cursor): cursor.execute('select 1 + 1;') return cursor.fetchone()[0] == 2 def error(): return False return sql(callback, error) def test_es(): try: return requests.get('http://localhost:9200').json()['tagline'] == 'You Know, for Search' except Exception: return False def sql(callback, error): try: with psycopg2.connect(host='localhost', port=5432, user='postgres', dbname='postgres') as conn: with conn.cursor() as cursor: return callback(cursor) except Exception: return error()
from os.path import abspath, dirname, join import sh import psycopg2 import requests import time PROJECT_FOLDER=dirname(dirname(abspath(__file__))) DOCKER_FOLDER=join(PROJECT_FOLDER, 'docker') def docker_compose(version, *args): sh.docker_compose('-f', 'docker/{version}/docker-compose.yml'.format(version=version), *args) def wait_for(condition): for i in xrange(120): if condition(): return True time.sleep(1) def test_pg(): try: return sql('select 1 + 1;')[0][0] == 2 except Exception: return False def test_es(): try: return requests.get('http://localhost:9200').json()['tagline'] == 'You Know, for Search' except Exception: return False def sql(statement): with psycopg2.connect(host='localhost', port=5432, user='postgres', dbname='postgres') as conn: with conn.cursor() as cursor: cursor.execute(statement) return cursor.fetchall()
Make the lib pretty complete
Make the lib pretty complete Should be able to replicate set up and tear down now
Python
mit
matthewfranglen/postgres-elasticsearch-fdw
--- +++ @@ -3,6 +3,7 @@ import sh import psycopg2 import requests +import time PROJECT_FOLDER=dirname(dirname(abspath(__file__))) DOCKER_FOLDER=join(PROJECT_FOLDER, 'docker') @@ -10,14 +11,17 @@ def docker_compose(version, *args): sh.docker_compose('-f', 'docker/{version}/docker-compose.yml'.format(version=version), *args) +def wait_for(condition): + for i in xrange(120): + if condition(): + return True + time.sleep(1) + def test_pg(): - def callback(cursor): - cursor.execute('select 1 + 1;') - return cursor.fetchone()[0] == 2 - def error(): + try: + return sql('select 1 + 1;')[0][0] == 2 + except Exception: return False - - return sql(callback, error) def test_es(): try: @@ -25,10 +29,8 @@ except Exception: return False -def sql(callback, error): - try: - with psycopg2.connect(host='localhost', port=5432, user='postgres', dbname='postgres') as conn: - with conn.cursor() as cursor: - return callback(cursor) - except Exception: - return error() +def sql(statement): + with psycopg2.connect(host='localhost', port=5432, user='postgres', dbname='postgres') as conn: + with conn.cursor() as cursor: + cursor.execute(statement) + return cursor.fetchall()
0b6cdcf91783e562d1da230e7658ba43b6ed5543
tests/test_unicode.py
tests/test_unicode.py
# coding: utf-8 import logging import os import shutil import sys import tempfile import unittest import six import fiona logging.basicConfig(stream=sys.stderr, level=logging.DEBUG) class UnicodePathTest(unittest.TestCase): def setUp(self): tempdir = tempfile.mkdtemp() self.dir = os.path.join(tempdir, 'français') shutil.copytree('docs/data/', self.dir) def tearDown(self): shutil.rmtree(self.dir) def test_unicode_path(self): path = self.dir + '/test_uk.shp' if sys.version_info < (3,): path = path.decode('utf-8') with fiona.open(path) as c: assert len(c) == 48 def test_unicode_path_layer(self): path = self.dir layer = 'test_uk' if sys.version_info < (3,): path = path.decode('utf-8') layer = layer.decode('utf-8') with fiona.open(path, layer=layer) as c: assert len(c) == 48 def test_utf8_path(self): path = self.dir + '/test_uk.shp' if sys.version_info < (3,): with fiona.open(path) as c: assert len(c) == 48
# coding: utf-8 import logging import os import shutil import sys import tempfile import unittest import six import fiona logging.basicConfig(stream=sys.stderr, level=logging.DEBUG) class UnicodePathTest(unittest.TestCase): def setUp(self): tempdir = tempfile.mkdtemp() self.dir = os.path.join(tempdir, 'français') shutil.copytree('docs/data/', self.dir) def tearDown(self): shutil.rmtree(os.path.dirname(self.dir)) def test_unicode_path(self): path = self.dir + '/test_uk.shp' if sys.version_info < (3,): path = path.decode('utf-8') with fiona.open(path) as c: assert len(c) == 48 def test_unicode_path_layer(self): path = self.dir layer = 'test_uk' if sys.version_info < (3,): path = path.decode('utf-8') layer = layer.decode('utf-8') with fiona.open(path, layer=layer) as c: assert len(c) == 48 def test_utf8_path(self): path = self.dir + '/test_uk.shp' if sys.version_info < (3,): with fiona.open(path) as c: assert len(c) == 48
Clean up parent temporary directory
TST: Clean up parent temporary directory
Python
bsd-3-clause
johanvdw/Fiona,rbuffat/Fiona,Toblerity/Fiona,perrygeo/Fiona,rbuffat/Fiona,Toblerity/Fiona,perrygeo/Fiona
--- +++ @@ -21,7 +21,7 @@ shutil.copytree('docs/data/', self.dir) def tearDown(self): - shutil.rmtree(self.dir) + shutil.rmtree(os.path.dirname(self.dir)) def test_unicode_path(self): path = self.dir + '/test_uk.shp'
25054586406024e082f9836884d5198ffa669f5b
models/ras_220_genes/build_ras_gene_network.py
models/ras_220_genes/build_ras_gene_network.py
from indra.tools.gene_network import GeneNetwork, grounding_filter import csv # STEP 0: Get gene list gene_list = [] # Get gene list from ras_pathway_proteins.csv with open('../../data/ras_pathway_proteins.csv') as f: csvreader = csv.reader(f, delimiter='\t') for row in csvreader: gene_list.append(row[0].strip()) gn = GeneNetwork(gene_list, 'ras_genes') stmts = gn.get_statements(filter=True) grounded_stmts = grounding_filter(stmts) results = gn.run_preassembly(grounded_stmts)
from indra.tools.gene_network import GeneNetwork, grounding_filter import csv import pickle # STEP 0: Get gene list gene_list = [] # Get gene list from ras_pathway_proteins.csv with open('../../data/ras_pathway_proteins.csv') as f: csvreader = csv.reader(f, delimiter='\t') for row in csvreader: gene_list.append(row[0].strip()) gn = GeneNetwork(gene_list, 'ras_genes') stmts = gn.get_statements(filter=True) grounded_stmts = grounding_filter(stmts) results = gn.run_preassembly(grounded_stmts) with open('ras_220_gn_stmts.pkl', 'wb') as f: pickle.dump(results, f)
Save the results of ras network
Save the results of ras network
Python
bsd-2-clause
bgyori/indra,sorgerlab/belpy,johnbachman/indra,johnbachman/belpy,johnbachman/belpy,sorgerlab/indra,johnbachman/belpy,pvtodorov/indra,jmuhlich/indra,bgyori/indra,sorgerlab/indra,pvtodorov/indra,johnbachman/indra,sorgerlab/belpy,sorgerlab/belpy,jmuhlich/indra,bgyori/indra,sorgerlab/indra,pvtodorov/indra,jmuhlich/indra,pvtodorov/indra,johnbachman/indra
--- +++ @@ -1,5 +1,6 @@ from indra.tools.gene_network import GeneNetwork, grounding_filter import csv +import pickle # STEP 0: Get gene list gene_list = [] @@ -13,4 +14,6 @@ stmts = gn.get_statements(filter=True) grounded_stmts = grounding_filter(stmts) results = gn.run_preassembly(grounded_stmts) +with open('ras_220_gn_stmts.pkl', 'wb') as f: + pickle.dump(results, f)
c98ab8807440e3cdbb98e11c53c7f246c35614fe
dedupe/convenience.py
dedupe/convenience.py
#!/usr/bin/python # -*- coding: utf-8 -*- """ Convenience functions for in memory deduplication """ import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' random_pairs = dedupe.core.randomPairs(len(data), sample_size) return tuple((data[k1], data[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
#!/usr/bin/python # -*- coding: utf-8 -*- """ Convenience functions for in memory deduplication """ import collections import dedupe.core def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' data_list = data.values() random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker): blocks = dedupe.core.OrderedDict({}) record_blocks = dedupe.core.OrderedDict({}) key_blocks = dedupe.core.OrderedDict({}) blocker.tfIdfBlocks(data_d.iteritems()) for (record_id, record) in data_d.iteritems(): for key in blocker((record_id, record)): blocks.setdefault(key, {}).update({record_id : record}) blocked_records = tuple(block for block in blocks.values()) return blocked_records
Change dataSample to generate indices of random pair using list of values
Change dataSample to generate indices of random pair using list of values
Python
mit
nmiranda/dedupe,01-/dedupe,neozhangthe1/dedupe,neozhangthe1/dedupe,nmiranda/dedupe,davidkunio/dedupe,dedupeio/dedupe,dedupeio/dedupe-examples,datamade/dedupe,tfmorris/dedupe,tfmorris/dedupe,davidkunio/dedupe,01-/dedupe,datamade/dedupe,pombredanne/dedupe,dedupeio/dedupe,pombredanne/dedupe
--- +++ @@ -13,10 +13,10 @@ def dataSample(data, sample_size): '''Randomly sample pairs of records from a data dictionary''' + data_list = data.values() + random_pairs = dedupe.core.randomPairs(len(data_list), sample_size) - random_pairs = dedupe.core.randomPairs(len(data), sample_size) - - return tuple((data[k1], data[k2]) for k1, k2 in random_pairs) + return tuple((data_list[k1], data_list[k2]) for k1, k2 in random_pairs) def blockData(data_d, blocker):
7733a84dc95d43070f476be42a3559b1a2a16ec0
dataset/print.py
dataset/print.py
import json with open('dataset_item.json') as dataset_file: dataset = json.load(dataset_file) for i in range(len(dataset)): if 'Continual' == dataset[i]['frequency']: print dataset[i]['name']
import json dataset = [] dataset_files = ['dataset_item.json'] for f in dataset_files: with open(f) as file: for line in file: dataset.append(json.loads(file)) for i in range(len(dataset)): if 'Continual' == dataset[i]['frequency']: print dataset[i]['name']
Update to line by line dataset JSON file parsing
Update to line by line dataset JSON file parsing New format under Feed Exports
Python
mit
MaxLikelihood/CODE
--- +++ @@ -1,7 +1,12 @@ import json -with open('dataset_item.json') as dataset_file: - dataset = json.load(dataset_file) +dataset = [] +dataset_files = ['dataset_item.json'] + +for f in dataset_files: + with open(f) as file: + for line in file: + dataset.append(json.loads(file)) for i in range(len(dataset)): if 'Continual' == dataset[i]['frequency']:
fe6891c949de75626396167a4aae78b276ed0223
pkg_resources/tests/test_markers.py
pkg_resources/tests/test_markers.py
try: import unitest.mock as mock except ImportError: import mock from pkg_resources import evaluate_marker @mock.patch('platform.python_version', return_value='2.7.10') def test_ordering(python_version_mock): assert evaluate_marker("python_full_version > '2.7.3'") is True
try: import unittest.mock as mock except ImportError: import mock from pkg_resources import evaluate_marker @mock.patch('platform.python_version', return_value='2.7.10') def test_ordering(python_version_mock): assert evaluate_marker("python_full_version > '2.7.3'") is True
Fix typo, correcting failures on late Pythons when mock is not already installed.
Fix typo, correcting failures on late Pythons when mock is not already installed.
Python
mit
pypa/setuptools,pypa/setuptools,pypa/setuptools
--- +++ @@ -1,5 +1,5 @@ try: - import unitest.mock as mock + import unittest.mock as mock except ImportError: import mock
2fd940a4c0eb047f2f5ac59fe04646e3e132e879
api_client/python/setup.py
api_client/python/setup.py
#!/usr/bin/env python # Copyright 2017 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """This is the setup file for the project.""" from setuptools import find_packages from setuptools import setup setup( name=u'timesketch-api-client', version=u'20170721', description=u'Timesketch', license=u'Apache License, Version 2.0', url=u'http://www.timesketch.org/', maintainer=u'Timesketch development team', maintainer_email=u'timesketch-dev@googlegroups.com', classifiers=[ u'Development Status :: 4 - Beta', u'Environment :: Console', u'Operating System :: OS Independent', u'Programming Language :: Python', ], packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=frozenset([ u'requests', u'BeautifulSoup', ]) )
#!/usr/bin/env python # Copyright 2017 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """This is the setup file for the project.""" from setuptools import find_packages from setuptools import setup setup( name=u'timesketch-api-client', version=u'20170815', description=u'Timesketch API client', license=u'Apache License, Version 2.0', url=u'http://www.timesketch.org/', maintainer=u'Timesketch development team', maintainer_email=u'timesketch-dev@googlegroups.com', classifiers=[ u'Development Status :: 4 - Beta', u'Environment :: Console', u'Operating System :: OS Independent', u'Programming Language :: Python', ], packages=find_packages(), include_package_data=True, zip_safe=False, install_requires=frozenset([ u'requests', u'BeautifulSoup', ]) )
Change version of API client
Change version of API client
Python
apache-2.0
google/timesketch,google/timesketch,google/timesketch,google/timesketch
--- +++ @@ -20,8 +20,8 @@ setup( name=u'timesketch-api-client', - version=u'20170721', - description=u'Timesketch', + version=u'20170815', + description=u'Timesketch API client', license=u'Apache License, Version 2.0', url=u'http://www.timesketch.org/', maintainer=u'Timesketch development team',
3de4665adae5f289fa896aa211ec32f72d956342
testproject/testproject/urls.py
testproject/testproject/urls.py
from django.conf import settings from django.conf.urls import patterns, include, url from django.conf.urls.static import static from django.contrib import admin from testproject import views urlpatterns = patterns('', # Examples: # url(r'^$', 'testproject.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^$', views.TestView.as_view(), name='home'), url(r'^groups-manager/', include('groups_manager.urls', namespace='groups_manager')), ) + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
from django.conf import settings from django.conf.urls import include, url from django.conf.urls.static import static from django.contrib import admin from testproject import views urlpatterns = [ # Examples: # url(r'^$', 'testproject.views.home', name='home'), # url(r'^blog/', include('blog.urls')), url(r'^admin/', include(admin.site.urls)), url(r'^$', views.TestView.as_view(), name='home'), url(r'^groups-manager/', include('groups_manager.urls', namespace='groups_manager')), ] + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
Remove use of deprecated patterns function
Remove use of deprecated patterns function
Python
mit
vittoriozamboni/django-groups-manager,vittoriozamboni/django-groups-manager
--- +++ @@ -1,11 +1,11 @@ from django.conf import settings -from django.conf.urls import patterns, include, url +from django.conf.urls import include, url from django.conf.urls.static import static from django.contrib import admin from testproject import views -urlpatterns = patterns('', +urlpatterns = [ # Examples: # url(r'^$', 'testproject.views.home', name='home'), # url(r'^blog/', include('blog.urls')), @@ -13,4 +13,4 @@ url(r'^admin/', include(admin.site.urls)), url(r'^$', views.TestView.as_view(), name='home'), url(r'^groups-manager/', include('groups_manager.urls', namespace='groups_manager')), -) + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT) +] + static(settings.STATIC_URL, document_root=settings.STATIC_ROOT)
a75d14ec1792404eadf4b23570c7d198839c97d2
day-02/solution.py
day-02/solution.py
from __future__ import print_function import fileinput from operator import mul from functools import reduce totalArea = 0 totalRibbon = 0 for line in fileinput.input(): parts = [int(i) for i in line.split('x')] parts.sort() sides = [parts[0] * parts[1], parts[0] * parts[2], parts[1] * parts[2]] totalArea += 2 * sum(sides) totalArea += min(sides) totalRibbon += 2 * (parts[0] + parts[1]) totalRibbon += reduce(mul, parts, 1) print(totalArea, "paper") print(totalRibbon, "ribbon")
from __future__ import print_function import fileinput from operator import mul from functools import reduce import itertools totalArea = 0 totalRibbon = 0 for line in fileinput.input(): parts = [int(i) for i in line.split('x')] parts.sort() sides = [x * y for x, y in itertools.combinations(parts, 2)] totalArea += 2 * sum(sides) + min(sides) totalRibbon += 2 * (parts[0] + parts[1]) totalRibbon += reduce(mul, parts, 1) print(totalArea, "paper") print(totalRibbon, "ribbon")
Use itertools for better readability.
Use itertools for better readability.
Python
mit
bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode,bertptrs/adventofcode
--- +++ @@ -2,6 +2,7 @@ import fileinput from operator import mul from functools import reduce +import itertools totalArea = 0 totalRibbon = 0 @@ -9,10 +10,9 @@ parts = [int(i) for i in line.split('x')] parts.sort() - sides = [parts[0] * parts[1], parts[0] * parts[2], parts[1] * parts[2]] - totalArea += 2 * sum(sides) + sides = [x * y for x, y in itertools.combinations(parts, 2)] + totalArea += 2 * sum(sides) + min(sides) - totalArea += min(sides) totalRibbon += 2 * (parts[0] + parts[1]) totalRibbon += reduce(mul, parts, 1)
68430e78313526a0b14b6d5abae810eb9c1ad53e
setup.py
setup.py
from setuptools import setup import os from appkit import __version__ data = list() for d in os.walk('appkit/'): if len(d[2]) > 0: path_list = [str.join('/', os.path.join(d[0], x).split('/')[1:]) for x in d[2]] data.extend(path_list) requires = ['flask', 'pygobject',] requires.append('beautifulsoup4') # v0_2_4 backward compatibility setup( name='appkit', version=__version__, description='Desktop application framework based on Webkit' + ' HTML5, CSS3, Javascript and Python', author='Nitipit Nontasuwan', author_email='nitipit@gmail.com', url='http://nitipit.github.com/appkit/', license='MIT', platforms=['Linux', ], keywords=['framework, html5, gnome, ui'], package_dir={'appkit': 'appkit'}, packages=['appkit'], package_data={'appkit': data}, install_requires=requires, )
from setuptools import setup import os from appkit import __version__ data = list() for d in os.walk('appkit/'): if len(d[2]) > 0: path_list = [str.join('/', os.path.join(d[0], x).split('/')[1:]) for x in d[2]] data.extend(path_list) requires = ['flask', 'pygobject',] requires.append('beautifulsoup4') # v0_2_4 backward compatibility setup( name='AppKit', version=__version__, description='Desktop application framework based on Webkit' + ' HTML5, CSS3, Javascript and Python', author='Nitipit Nontasuwan', author_email='nitipit@gmail.com', url='http://nitipit.github.com/appkit/', license='MIT', platforms=['Linux', ], keywords=['framework, html5, gnome, ui'], package_dir={'appkit': 'appkit'}, packages=['appkit'], package_data={'appkit': data}, install_requires=requires, )
Fix package name to match PyPI
Fix package name to match PyPI
Python
mit
nitipit/appkit
--- +++ @@ -13,7 +13,7 @@ requires.append('beautifulsoup4') # v0_2_4 backward compatibility setup( - name='appkit', + name='AppKit', version=__version__, description='Desktop application framework based on Webkit' + ' HTML5, CSS3, Javascript and Python',
c07bacb73eec4b963ec53c067f23385dad246fb6
setup.py
setup.py
from distutils.core import setup setup(name='zencoder', version='0.5.2', description='Integration library for Zencoder', author='Alex Schworer', author_email='alex.schworer@gmail.com', url='http://github.com/schworer/zencoder-py', license="MIT License", install_requires=['httplib2'], packages=['zencoder'] )
from distutils.core import setup setup(name='zencoder', version='0.5.2', description='Integration library for Zencoder', author='Alex Schworer', author_email='alex.schworer@gmail.com', url='http://github.com/schworer/zencoder-py', license="MIT License", install_requires=['httplib2'], packages=['zencoder'], platforms='any', classifiers=[ 'Development Status :: 4 - Beta', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2.5', 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Topic :: Software Development :: Libraries :: Python Modules' ] )
Add classifiers to zencoder-py package
Add classifiers to zencoder-py package
Python
mit
zencoder/zencoder-py
--- +++ @@ -9,6 +9,18 @@ url='http://github.com/schworer/zencoder-py', license="MIT License", install_requires=['httplib2'], - packages=['zencoder'] + packages=['zencoder'], + platforms='any', + classifiers=[ + 'Development Status :: 4 - Beta', + 'Intended Audience :: Developers', + 'License :: OSI Approved :: MIT License', + 'Operating System :: OS Independent', + 'Programming Language :: Python', + 'Programming Language :: Python :: 2.5', + 'Programming Language :: Python :: 2.6', + 'Programming Language :: Python :: 2.7', + 'Topic :: Software Development :: Libraries :: Python Modules' + ] )
2eea58a64c57c5a66c13220a66b92c4dc9f7fdb3
setup.py
setup.py
import sys from setuptools import setup, find_packages py26_dependency = [] if sys.version_info.major == 2 and sys.version_info.minor < 7: py26_dependency = ["argparse >= 1.2.1"] setup( name='dataset', version='0.3.14', description="Toolkit for Python-based data processing.", long_description="", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.3' ], keywords='sql sqlalchemy etl loading utility', author='Friedrich Lindenberg, Gregor Aisch', author_email='info@okfn.org', url='http://github.com/pudo/dataset', license='MIT', packages=find_packages(exclude=['ez_setup', 'examples', 'tests']), namespace_packages=[], include_package_data=False, zip_safe=False, install_requires=[ 'sqlalchemy >= 0.8.1', 'alembic >= 0.6.1', 'python-slugify >= 0.0.6', "PyYAML >= 3.10" ] + py26_dependency, tests_require=[], entry_points={ 'console_scripts': [ 'datafreeze = dataset.freeze.app:main', ] } )
import sys from setuptools import setup, find_packages py26_dependency = [] if sys.version_info <= (2, 6): py26_dependency = ["argparse >= 1.2.1"] setup( name='dataset', version='0.3.14', description="Toolkit for Python-based data processing.", long_description="", classifiers=[ "Development Status :: 3 - Alpha", "Intended Audience :: Developers", "License :: OSI Approved :: MIT License", "Operating System :: OS Independent", 'Programming Language :: Python :: 2.6', 'Programming Language :: Python :: 2.7', 'Programming Language :: Python :: 3.3' ], keywords='sql sqlalchemy etl loading utility', author='Friedrich Lindenberg, Gregor Aisch', author_email='info@okfn.org', url='http://github.com/pudo/dataset', license='MIT', packages=find_packages(exclude=['ez_setup', 'examples', 'tests']), namespace_packages=[], include_package_data=False, zip_safe=False, install_requires=[ 'sqlalchemy >= 0.8.1', 'alembic >= 0.6.1', 'python-slugify >= 0.0.6', "PyYAML >= 3.10" ] + py26_dependency, tests_require=[], entry_points={ 'console_scripts': [ 'datafreeze = dataset.freeze.app:main', ] } )
Fix Python version detection for Python 2.6
Fix Python version detection for Python 2.6
Python
mit
reubano/dataset,vguzmanp/dataset,pudo/dataset,askebos/dataset,stefanw/dataset,twds/dataset,saimn/dataset
--- +++ @@ -3,7 +3,7 @@ py26_dependency = [] -if sys.version_info.major == 2 and sys.version_info.minor < 7: +if sys.version_info <= (2, 6): py26_dependency = ["argparse >= 1.2.1"] setup(
dfca8ac68d69e533b954462094890faf0e723891
autopoke.py
autopoke.py
#!/bin/env python from selenium import webdriver from selenium.common.exceptions import StaleElementReferenceException from time import sleep from getpass import getpass if __name__ == '__main__': driver = webdriver.phantomjs.webdriver.WebDriver() driver.get('https://facebook.com') driver.find_element_by_id('email').send_keys(input('Email: ')) driver.find_element_by_id('pass').send_keys(getpass()) driver.find_element_by_id('loginbutton').click() driver.get('https://facebook.com/pokes/') assert "Forgot password?" not in driver.page_source c = 0 while True: try: for i in driver.find_elements_by_link_text("Poke Back"): i.click() c += 1 print("Clicked so far: " + str(c)) except StaleElementReferenceException: print("Found exception, reloading page") driver.get('https://facebook.com/pokes/') sleep(0.5)
#!/bin/env python from selenium import webdriver from selenium.common.exceptions import StaleElementReferenceException from time import sleep from getpass import getpass if __name__ == '__main__': driver = webdriver.phantomjs.webdriver.WebDriver() driver.get('https://facebook.com') driver.find_element_by_id('email').send_keys(input('Email: ')) driver.find_element_by_id('pass').send_keys(getpass()) driver.find_element_by_id('loginbutton').click() driver.get('https://facebook.com/pokes/') assert "Forgot password?" not in driver.page_source c = 0 c2 = 0 while True: try: for i in driver.find_elements_by_link_text("Poke Back"): i.click() c += 1 c2 = 0 print("Clicked so far: " + str(c)) except StaleElementReferenceException: print("Found exception, reloading page") driver.get('https://facebook.com/pokes/') c2 += 1 if c2 % 121 == 0: print("No pokes in last minute. Reloading") driver.get('https://facebook.com/pokes/') sleep(0.5)
Fix bug where page stops updating by forcing it to reload after a minute of no activity
Fix bug where page stops updating by forcing it to reload after a minute of no activity
Python
mit
matthewbentley/autopoke
--- +++ @@ -15,14 +15,20 @@ assert "Forgot password?" not in driver.page_source c = 0 + c2 = 0 while True: try: for i in driver.find_elements_by_link_text("Poke Back"): i.click() c += 1 + c2 = 0 print("Clicked so far: " + str(c)) except StaleElementReferenceException: print("Found exception, reloading page") driver.get('https://facebook.com/pokes/') + c2 += 1 + if c2 % 121 == 0: + print("No pokes in last minute. Reloading") + driver.get('https://facebook.com/pokes/') sleep(0.5)
7cb65a86a1b0fe8a739fb81fb8a66f3d205142cb
setup.py
setup.py
# Licensed under a 3-clause BSD style license - see LICENSE.rst from chandra_aca import __version__ from setuptools import setup try: from testr.setup_helper import cmdclass except ImportError: cmdclass = {} setup(name='chandra_aca', author='Jean Connelly, Tom Aldcroft', description='Chandra Aspect Camera Tools', author_email='jconnelly@cfa.harvard.edu', version=__version__, zip_safe=False, packages=['chandra_aca', 'chandra_aca.tests'], package_data={'chandra_aca.tests': ['data/*.txt', 'data/*.dat'], 'chandra_aca': ['data/*.dat', 'data/star_probs/*.fits.gz']}, tests_require=['pytest'], cmdclass=cmdclass, )
# Licensed under a 3-clause BSD style license - see LICENSE.rst from chandra_aca import __version__ from setuptools import setup try: from testr.setup_helper import cmdclass except ImportError: cmdclass = {} setup(name='chandra_aca', author='Jean Connelly, Tom Aldcroft', description='Chandra Aspect Camera Tools', author_email='jconnelly@cfa.harvard.edu', version=__version__, zip_safe=False, packages=['chandra_aca', 'chandra_aca.tests'], package_data={'chandra_aca.tests': ['data/*.txt', 'data/*.dat'], 'chandra_aca': ['data/*.dat', 'data/*.fits.gz', 'data/star_probs/*.fits.gz']}, tests_require=['pytest'], cmdclass=cmdclass, )
Add missing package data for flickering
Add missing package data for flickering
Python
bsd-2-clause
sot/chandra_aca,sot/chandra_aca
--- +++ @@ -16,7 +16,8 @@ zip_safe=False, packages=['chandra_aca', 'chandra_aca.tests'], package_data={'chandra_aca.tests': ['data/*.txt', 'data/*.dat'], - 'chandra_aca': ['data/*.dat', 'data/star_probs/*.fits.gz']}, + 'chandra_aca': ['data/*.dat', 'data/*.fits.gz', + 'data/star_probs/*.fits.gz']}, tests_require=['pytest'], cmdclass=cmdclass, )
9668fdcfe67bea8d8c84dedd1e5f9e2646474e76
setup.py
setup.py
from distutils.core import setup with open('README.md') as f: long_description = f.read() setup( name = 'dont_argue', packages = ['dont_argue'], version = '0.1.1', description = 'Dead-simple command line argument parsing', long_description=long_description, author = 'Chris Penner', author_email = 'christopher.penner@gmail.com', url = 'https://github.com/ChrisPenner/dont-argue', download_url = 'https://github.com/ChrisPenner/dont-argue/releases/tag/0.1.1', license = 'MIT', keywords = ['command line', 'argument', 'parsing', 'argparse'], classifiers = [ 'Development Status :: 3 - Alpha', 'Topic :: Software Development :: Libraries', # Pick your license as you wish (should match "license" above) 'License :: OSI Approved :: MIT License', # Specify the Python versions you support here. In particular, ensure # that you indicate whether you support Python 2, Python 3 or both. 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ], )
from distutils.core import setup with open('README.md') as f: long_description = f.read() setup( name = 'dont_argue', packages = ['dont_argue'], version = '0.1.1', description = 'Dead-simple command line argument parsing', long_description=long_description, author = 'Chris Penner', author_email = 'christopher.penner@gmail.com', url = 'https://github.com/ChrisPenner/dont-argue', download_url = 'https://github.com/ChrisPenner/dont-argue/releases/tag/v0.1.1', license = 'MIT', keywords = ['command line', 'argument', 'parsing', 'argparse'], classifiers = [ 'Development Status :: 3 - Alpha', 'Topic :: Software Development :: Libraries', # Pick your license as you wish (should match "license" above) 'License :: OSI Approved :: MIT License', # Specify the Python versions you support here. In particular, ensure # that you indicate whether you support Python 2, Python 3 or both. 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', ], )
Change release tagging convention to vX.Y.Z
Change release tagging convention to vX.Y.Z
Python
mit
ChrisPenner/dont-argue
--- +++ @@ -12,7 +12,7 @@ author = 'Chris Penner', author_email = 'christopher.penner@gmail.com', url = 'https://github.com/ChrisPenner/dont-argue', - download_url = 'https://github.com/ChrisPenner/dont-argue/releases/tag/0.1.1', + download_url = 'https://github.com/ChrisPenner/dont-argue/releases/tag/v0.1.1', license = 'MIT', keywords = ['command line', 'argument', 'parsing', 'argparse'], classifiers = [
2af4061d0add17e54c7ded45e34177efdffbc7db
setup.py
setup.py
# -*- coding: utf-8 -*- import sys if sys.version_info[0] != 2: sys.exit("Sorry, Python 3 is not supported yet") from setuptools import setup, find_packages install_requires = [ 'pytz', 'requests', 'lxml', 'python-dateutil', 'parsedatetime', 'cached-property', 'click', 'enum34', # backported versions from Python3 'pathlib', 'configparser', 'zope.interface', 'attrs', ] console_scripts = [ 'poker = poker.commands:poker', ] classifiers = [ "Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", ] setup( name='poker', version='0.23.1', description='Poker Framework', long_description=open('README.rst', 'r').read().decode('utf-8'), classifiers=classifiers, keywords='poker', author=u'Kiss György', author_email="kissgyorgy@me.com", url="https://github.com/pokerregion/poker", license="MIT", packages=find_packages(), install_requires=install_requires, entry_points={'console_scripts': console_scripts}, tests_require=['pytest', 'coverage', 'coveralls'], )
# -*- coding: utf-8 -*- from pathlib import Path from setuptools import setup, find_packages install_requires = [ 'pytz', 'requests', 'lxml', 'python-dateutil', 'parsedatetime', 'cached-property', 'click', 'enum34', # backported versions from Python3 'pathlib', 'configparser', 'zope.interface', 'attrs', ] console_scripts = [ 'poker = poker.commands:poker', ] classifiers = [ "Intended Audience :: Developers", "Programming Language :: Python", "Programming Language :: Python :: 2", "Programming Language :: Python :: 2.7", ] setup( name='poker', version='0.23.1', description='Poker Framework', long_description=Path('README.rst').read_text(), classifiers=classifiers, keywords='poker', author=u'Kiss György', author_email="kissgyorgy@me.com", url="https://github.com/pokerregion/poker", license="MIT", packages=find_packages(), install_requires=install_requires, entry_points={'console_scripts': console_scripts}, tests_require=['pytest', 'coverage', 'coveralls'], )
Make the package installable with python3
Make the package installable with python3
Python
mit
pokerregion/poker
--- +++ @@ -1,9 +1,5 @@ # -*- coding: utf-8 -*- - -import sys -if sys.version_info[0] != 2: - sys.exit("Sorry, Python 3 is not supported yet") - +from pathlib import Path from setuptools import setup, find_packages @@ -40,7 +36,7 @@ name='poker', version='0.23.1', description='Poker Framework', - long_description=open('README.rst', 'r').read().decode('utf-8'), + long_description=Path('README.rst').read_text(), classifiers=classifiers, keywords='poker', author=u'Kiss György',
27c6df9b0e936ce4fb173ec64230931ffe0719c7
querylist/querylist.py
querylist/querylist.py
from betterdict import BetterDict class QueryList(list): def __init__(self, data=None, wrapper=BetterDict): """Create a QueryList from an iterable and a wrapper object.""" self._wrapper = wrapper self.src_data = data # Wrap our src_data with wrapper converted_data = self._convert_iterable(data) if data else [] super(QueryList, self).__init__(converted_data) def _convert_iterable(self, iterable): # Return original if _wrapper isn't callable if not callable(self._wrapper): return iterable return [self._wrapper(x) for x in iterable]
from betterdict import BetterDict class QueryList(list): def __init__(self, data=None, wrapper=BetterDict): """Create a QueryList from an iterable and a wrapper object.""" self._wrapper = wrapper self.src_data = data # Wrap our src_data with wrapper converted_data = self._convert_iterable(data) if data else [] super(QueryList, self).__init__(converted_data) def _convert_iterable(self, iterable): """Converts elements returned by an iterable into instances of self._wrapper """ # Return original if _wrapper isn't callable if not callable(self._wrapper): return iterable return [self._wrapper(x) for x in iterable]
Add missing docs tring to QueryList._convert_iterable()
Add missing docs tring to QueryList._convert_iterable()
Python
mit
zoidbergwill/querylist,thomasw/querylist
--- +++ @@ -13,6 +13,10 @@ super(QueryList, self).__init__(converted_data) def _convert_iterable(self, iterable): + """Converts elements returned by an iterable into instances of + self._wrapper + + """ # Return original if _wrapper isn't callable if not callable(self._wrapper): return iterable
3d9911a6a2d24631b21850da6fe8f04787465b9e
example-flask-python3.6-index/app/main.py
example-flask-python3.6-index/app/main.py
from flask import Flask, send_file app = Flask(__name__) @app.route("/hello") def hello(): return "Hello World from Flask in a uWSGI Nginx Docker container with \ Python 3.6 (from the example template)" @app.route("/") def main(): return send_file('./static/index.html') # Everything not declared before (not a Flask route / API endpoint)... @app.route('/<path:path>') def route_frontend(path): # ...could be a static file needed by the front end that # doesn't use the `static` path (like in `<script src="bundle.js">`) file_path = './static/' + path if os.path.isfile(file_path): return send_file(file_path) # ...or should be handled by the SPA's "router" in front end else: return send_file('./static/index.html') if __name__ == "__main__": app.run(host='0.0.0.0', debug=True, port=80)
import os from flask import Flask, send_file app = Flask(__name__) @app.route("/hello") def hello(): return "Hello World from Flask in a uWSGI Nginx Docker container with \ Python 3.6 (from the example template)" @app.route("/") def main(): return send_file('./static/index.html') # Everything not declared before (not a Flask route / API endpoint)... @app.route('/<path:path>') def route_frontend(path): # ...could be a static file needed by the front end that # doesn't use the `static` path (like in `<script src="bundle.js">`) file_path = './static/' + path if os.path.isfile(file_path): return send_file(file_path) # ...or should be handled by the SPA's "router" in front end else: return send_file('./static/index.html') if __name__ == "__main__": app.run(host='0.0.0.0', debug=True, port=80)
Fix not imported os and format code
Fix not imported os and format code
Python
apache-2.0
tiangolo/uwsgi-nginx-flask-docker,tiangolo/uwsgi-nginx-flask-docker,tiangolo/uwsgi-nginx-flask-docker
--- +++ @@ -1,14 +1,19 @@ +import os + from flask import Flask, send_file app = Flask(__name__) + @app.route("/hello") def hello(): return "Hello World from Flask in a uWSGI Nginx Docker container with \ Python 3.6 (from the example template)" + @app.route("/") def main(): return send_file('./static/index.html') + # Everything not declared before (not a Flask route / API endpoint)... @app.route('/<path:path>') @@ -22,5 +27,6 @@ else: return send_file('./static/index.html') + if __name__ == "__main__": app.run(host='0.0.0.0', debug=True, port=80)
31d0cd541980ef6bf15d3a29b68cc0cc994c28a4
packs/st2cd/actions/kvstore.py
packs/st2cd/actions/kvstore.py
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = KeyValuePair() instance.id = client.keys.get_by_name(key).name instance.name = key instance.value = value try: kvstore = getattr(client.keys, action) kvp = kvstore(instance) except Exception as e: raise if action == 'delete': return kvp else: return kvp.serialize()
from st2actions.runners.pythonrunner import Action from st2client.client import Client from st2client.models.datastore import KeyValuePair class KVPAction(Action): def run(self, key, action, st2host='localhost', value=""): st2_endpoints = { 'action': "http://%s:9101" % st2host, 'reactor': "http://%s:9102" % st2host, 'datastore': "http://%s:9103" % st2host } try: client = Client(st2_endpoints) except Exception as e: return e if action == 'get': kvp = client.keys.get_by_name(key) if not kvp: raise Exception('Key error with %s.' % key) return kvp.value else: instance = client.keys.get_by_name(key) or KeyValuePair() instance.id = key instance.name = key instance.value = value kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp else: return kvp.serialize()
Fix create action for key value pair
Fix create action for key value pair
Python
apache-2.0
StackStorm/st2incubator,pinterb/st2incubator,pinterb/st2incubator,pinterb/st2incubator,StackStorm/st2incubator
--- +++ @@ -25,16 +25,12 @@ return kvp.value else: - instance = KeyValuePair() - instance.id = client.keys.get_by_name(key).name + instance = client.keys.get_by_name(key) or KeyValuePair() + instance.id = key instance.name = key instance.value = value - try: - kvstore = getattr(client.keys, action) - kvp = kvstore(instance) - except Exception as e: - raise + kvp = client.keys.update(instance) if action in ['create', 'update'] else None if action == 'delete': return kvp
c2272f7d23645932b72617b1b20e5ddd86267ef0
app/letter_branding/letter_branding_rest.py
app/letter_branding/letter_branding_rest.py
from flask import Blueprint from app.errors import register_errors email_branding_blueprint = Blueprint('letter_branding', __name__, url_prefix='letter-branding') register_errors(email_branding_blueprint)
from flask import Blueprint from app.errors import register_errors email_branding_blueprint = Blueprint('letter_branding', __name__, url_prefix='letter-branding') register_errors(email_branding_blueprint)
Fix mock in unit test. Not sure why it passed locally with previous mock.
Fix mock in unit test. Not sure why it passed locally with previous mock.
Python
mit
alphagov/notifications-api,alphagov/notifications-api
a0ff3446a177d11268c49b37c03f0d495341fe81
banana/maya/extensions/OpenMaya/__init__.py
banana/maya/extensions/OpenMaya/__init__.py
""" banana.maya.extensions.OpenMaya ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ OpenMaya extensions. :copyright: Copyright 2014 by Christopher Crouzet. :license: MIT, see LICENSE for details. """ __all__ = [ 'iterators', 'MDagPath', 'MFileIO', 'MFnDagNode', 'MFnDependencyNode', 'MFnTransform', 'MObject' ]
""" banana.maya.extensions.OpenMaya ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ OpenMaya extensions. :copyright: Copyright 2014 by Christopher Crouzet. :license: MIT, see LICENSE for details. """ __all__ = [ 'iterators', 'MDagPath', 'MFileIO', 'MFnDagNode', 'MFnDependencyNode', 'MFnTransform', 'MGlobal', 'MObject' ]
Add the missing `MGlobal` module to the `__all__` attribute.
Add the missing `MGlobal` module to the `__all__` attribute.
Python
mit
christophercrouzet/bana,christophercrouzet/banana.maya
--- +++ @@ -15,5 +15,6 @@ 'MFnDagNode', 'MFnDependencyNode', 'MFnTransform', + 'MGlobal', 'MObject' ]
5868f46a05ef19862ea81b4e402851a3e40ceeff
events/serializers.py
events/serializers.py
from .models import Event, EventActivity from employees.serializers import LocationSerializer from rest_framework import serializers class EventSerializer(serializers.ModelSerializer): location = LocationSerializer() class Meta(object): model = Event depth = 1 fields = ("pk", "name", "image", "datetime", "address", "description", "is_active", "is_upcoming", "location") class EventSimpleSerializer(serializers.Serializer): pk = serializers.IntegerField() name = serializers.CharField(max_length=100) image = serializers.CharField(allow_blank=True, required=False) datetime = serializers.DateTimeField(required=False) address = serializers.CharField(allow_blank=True, required=False) description = serializers.CharField(allow_blank=True, required=False) is_registered = serializers.BooleanField() class EventActivitySerializer(serializers.ModelSerializer): class Meta(object): model = EventActivity
from .models import Event, EventActivity from employees.serializers import LocationSerializer from rest_framework import serializers class EventSerializer(serializers.ModelSerializer): location = LocationSerializer() class Meta(object): model = Event depth = 1 fields = ("pk", "name", "image", "datetime", "address", "description", "is_active", "location") class EventSimpleSerializer(serializers.Serializer): pk = serializers.IntegerField() name = serializers.CharField(max_length=100) image = serializers.CharField(allow_blank=True, required=False) datetime = serializers.DateTimeField(required=False) address = serializers.CharField(allow_blank=True, required=False) description = serializers.CharField(allow_blank=True, required=False) is_registered = serializers.BooleanField() class EventActivitySerializer(serializers.ModelSerializer): class Meta(object): model = EventActivity fields = ("pk", "datetime", "text", "event")
Remove is_upcoming field from Event response and Add explicit fields to EventActivity serializer
Remove is_upcoming field from Event response and Add explicit fields to EventActivity serializer
Python
apache-2.0
belatrix/BackendAllStars
--- +++ @@ -9,7 +9,7 @@ class Meta(object): model = Event depth = 1 - fields = ("pk", "name", "image", "datetime", "address", "description", "is_active", "is_upcoming", "location") + fields = ("pk", "name", "image", "datetime", "address", "description", "is_active", "location") class EventSimpleSerializer(serializers.Serializer): @@ -25,3 +25,4 @@ class EventActivitySerializer(serializers.ModelSerializer): class Meta(object): model = EventActivity + fields = ("pk", "datetime", "text", "event")
f9f41ec4f27ba5fd19ca82d4c04b13bed6627d23
app/PRESUBMIT.py
app/PRESUBMIT.py
#!/usr/bin/python # Copyright (c) 2009 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Makes sure that the app/ code is cpplint clean.""" INCLUDE_CPP_FILES_ONLY = ( r'.*\.cc$', r'.*\.h$' ) EXCLUDE = ( # Autogenerated window resources files are off limits r'.*resource.h$', ) def CheckChangeOnUpload(input_api, output_api): results = [] black_list = input_api.DEFAULT_BLACK_LIST + EXCLUDE sources = lambda x: input_api.FilterSourceFile( x, white_list=INCLUDE_CPP_FILES_ONLY, black_list=black_list) results.extend(input_api.canned_checks.CheckChangeLintsClean( input_api, output_api, sources)) return results
#!/usr/bin/python # Copyright (c) 2009 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. """Makes sure that the app/ code is cpplint clean.""" INCLUDE_CPP_FILES_ONLY = ( r'.*\.cc$', r'.*\.h$' ) EXCLUDE = ( # Autogenerated window resources files are off limits r'.*resource.h$', ) def CheckChangeOnUpload(input_api, output_api): results = [] black_list = input_api.DEFAULT_BLACK_LIST + EXCLUDE sources = lambda x: input_api.FilterSourceFile( x, white_list=INCLUDE_CPP_FILES_ONLY, black_list=black_list) results.extend(input_api.canned_checks.CheckChangeLintsClean( input_api, output_api, sources)) return results def GetPreferredTrySlaves(): return ['win', 'linux', 'linux_view', 'linux_chromeos', 'mac']
Make all changes to app/ run on all trybot platforms, not just the big three. Anyone who's changing a header here may break the chromeos build.
Make all changes to app/ run on all trybot platforms, not just the big three. Anyone who's changing a header here may break the chromeos build. BUG=none TEST=none Review URL: http://codereview.chromium.org/2838027 git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@51000 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
rogerwang/chromium,Fireblend/chromium-crosswalk,dednal/chromium.src,TheTypoMaster/chromium-crosswalk,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,ChromiumWebApps/chromium,dushu1203/chromium.src,ChromiumWebApps/chromium,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,Fireblend/chromium-crosswalk,zcbenz/cefode-chromium,hgl888/chromium-crosswalk-efl,littlstar/chromium.src,anirudhSK/chromium,hujiajie/pa-chromium,hujiajie/pa-chromium,fujunwei/chromium-crosswalk,chuan9/chromium-crosswalk,Chilledheart/chromium,pozdnyakov/chromium-crosswalk,rogerwang/chromium,jaruba/chromium.src,littlstar/chromium.src,dushu1203/chromium.src,mohamed--abdel-maksoud/chromium.src,Pluto-tv/chromium-crosswalk,krieger-od/nwjs_chromium.src,pozdnyakov/chromium-crosswalk,robclark/chromium,keishi/chromium,nacl-webkit/chrome_deps,M4sse/chromium.src,M4sse/chromium.src,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk-efl,Jonekee/chromium.src,markYoungH/chromium.src,pozdnyakov/chromium-crosswalk,junmin-zhu/chromium-rivertrail,PeterWangIntel/chromium-crosswalk,chuan9/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,dednal/chromium.src,Just-D/chromium-1,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,ondra-novak/chromium.src,junmin-zhu/chromium-rivertrail,dednal/chromium.src,jaruba/chromium.src,zcbenz/cefode-chromium,keishi/chromium,zcbenz/cefode-chromium,markYoungH/chromium.src,patrickm/chromium.src,fujunwei/chromium-crosswalk,krieger-od/nwjs_chromium.src,ltilve/chromium,anirudhSK/chromium,keishi/chromium,bright-sparks/chromium-spacewalk,anirudhSK/chromium,nacl-webkit/chrome_deps,nacl-webkit/chrome_deps,axinging/chromium-crosswalk,jaruba/chromium.src,jaruba/chromium.src,Chilledheart/chromium,chuan9/chromium-crosswalk,ltilve/chromium,dushu1203/chromium.src,patrickm/chromium.src,ChromiumWebApps/chromium,zcbenz/cefode-chromium,ondra-novak/chromium.src,pozdnyakov/chromium-crosswalk,M4sse/chromium.src,Just-D/chromium-1,timopulkkinen/BubbleFish,junmin-zhu/chromium-rivertrail,hgl888/chromium-crosswalk-efl,jaruba/chromium.src,crosswalk-project/chromium-crosswalk-efl,nacl-webkit/chrome_deps,Jonekee/chromium.src,anirudhSK/chromium,krieger-od/nwjs_chromium.src,hujiajie/pa-chromium,krieger-od/nwjs_chromium.src,robclark/chromium,anirudhSK/chromium,robclark/chromium,bright-sparks/chromium-spacewalk,fujunwei/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,chuan9/chromium-crosswalk,Pluto-tv/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,jaruba/chromium.src,robclark/chromium,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,zcbenz/cefode-chromium,mohamed--abdel-maksoud/chromium.src,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,ChromiumWebApps/chromium,hgl888/chromium-crosswalk-efl,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,timopulkkinen/BubbleFish,dednal/chromium.src,M4sse/chromium.src,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,jaruba/chromium.src,fujunwei/chromium-crosswalk,hgl888/chromium-crosswalk,axinging/chromium-crosswalk,anirudhSK/chromium,TheTypoMaster/chromium-crosswalk,rogerwang/chromium,mohamed--abdel-maksoud/chromium.src,Jonekee/chromium.src,hujiajie/pa-chromium,dushu1203/chromium.src,Chilledheart/chromium,M4sse/chromium.src,dushu1203/chromium.src,Just-D/chromium-1,dednal/chromium.src,littlstar/chromium.src,axinging/chromium-crosswalk,Jonekee/chromium.src,rogerwang/chromium,keishi/chromium,M4sse/chromium.src,chuan9/chromium-crosswalk,Fireblend/chromium-crosswalk,junmin-zhu/chromium-rivertrail,ChromiumWebApps/chromium,Jonekee/chromium.src,hgl888/chromium-crosswalk,zcbenz/cefode-chromium,Chilledheart/chromium,Just-D/chromium-1,axinging/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,pozdnyakov/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,keishi/chromium,hgl888/chromium-crosswalk,hujiajie/pa-chromium,krieger-od/nwjs_chromium.src,patrickm/chromium.src,Chilledheart/chromium,mogoweb/chromium-crosswalk,pozdnyakov/chromium-crosswalk,ltilve/chromium,mogoweb/chromium-crosswalk,pozdnyakov/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,Fireblend/chromium-crosswalk,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,pozdnyakov/chromium-crosswalk,mogoweb/chromium-crosswalk,markYoungH/chromium.src,mogoweb/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,keishi/chromium,fujunwei/chromium-crosswalk,hgl888/chromium-crosswalk-efl,ChromiumWebApps/chromium,mohamed--abdel-maksoud/chromium.src,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk,mogoweb/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,axinging/chromium-crosswalk,robclark/chromium,Fireblend/chromium-crosswalk,mogoweb/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ondra-novak/chromium.src,krieger-od/nwjs_chromium.src,crosswalk-project/chromium-crosswalk-efl,krieger-od/nwjs_chromium.src,mohamed--abdel-maksoud/chromium.src,markYoungH/chromium.src,patrickm/chromium.src,littlstar/chromium.src,Just-D/chromium-1,patrickm/chromium.src,Jonekee/chromium.src,markYoungH/chromium.src,chuan9/chromium-crosswalk,littlstar/chromium.src,ondra-novak/chromium.src,robclark/chromium,ltilve/chromium,jaruba/chromium.src,Jonekee/chromium.src,hgl888/chromium-crosswalk,dednal/chromium.src,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk,Pluto-tv/chromium-crosswalk,ondra-novak/chromium.src,ondra-novak/chromium.src,mohamed--abdel-maksoud/chromium.src,PeterWangIntel/chromium-crosswalk,hujiajie/pa-chromium,keishi/chromium,Fireblend/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,nacl-webkit/chrome_deps,patrickm/chromium.src,TheTypoMaster/chromium-crosswalk,markYoungH/chromium.src,rogerwang/chromium,pozdnyakov/chromium-crosswalk,jaruba/chromium.src,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,hgl888/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,zcbenz/cefode-chromium,markYoungH/chromium.src,robclark/chromium,M4sse/chromium.src,Just-D/chromium-1,hujiajie/pa-chromium,timopulkkinen/BubbleFish,dednal/chromium.src,mogoweb/chromium-crosswalk,junmin-zhu/chromium-rivertrail,zcbenz/cefode-chromium,M4sse/chromium.src,robclark/chromium,ChromiumWebApps/chromium,jaruba/chromium.src,ChromiumWebApps/chromium,dednal/chromium.src,dushu1203/chromium.src,Just-D/chromium-1,ondra-novak/chromium.src,crosswalk-project/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,anirudhSK/chromium,crosswalk-project/chromium-crosswalk-efl,keishi/chromium,fujunwei/chromium-crosswalk,littlstar/chromium.src,bright-sparks/chromium-spacewalk,Fireblend/chromium-crosswalk,dushu1203/chromium.src,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,timopulkkinen/BubbleFish,Fireblend/chromium-crosswalk,hujiajie/pa-chromium,timopulkkinen/BubbleFish,fujunwei/chromium-crosswalk,fujunwei/chromium-crosswalk,markYoungH/chromium.src,anirudhSK/chromium,chuan9/chromium-crosswalk,junmin-zhu/chromium-rivertrail,ltilve/chromium,hgl888/chromium-crosswalk,ltilve/chromium,junmin-zhu/chromium-rivertrail,patrickm/chromium.src,Jonekee/chromium.src,nacl-webkit/chrome_deps,ltilve/chromium,zcbenz/cefode-chromium,timopulkkinen/BubbleFish,timopulkkinen/BubbleFish,hgl888/chromium-crosswalk-efl,markYoungH/chromium.src,dushu1203/chromium.src,bright-sparks/chromium-spacewalk,Just-D/chromium-1,chuan9/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,hujiajie/pa-chromium,dushu1203/chromium.src,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,Pluto-tv/chromium-crosswalk,Jonekee/chromium.src,hujiajie/pa-chromium,Chilledheart/chromium,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,bright-sparks/chromium-spacewalk,rogerwang/chromium,Pluto-tv/chromium-crosswalk,anirudhSK/chromium,ondra-novak/chromium.src,fujunwei/chromium-crosswalk,nacl-webkit/chrome_deps,Jonekee/chromium.src,zcbenz/cefode-chromium,dednal/chromium.src,Chilledheart/chromium,Pluto-tv/chromium-crosswalk,Pluto-tv/chromium-crosswalk,timopulkkinen/BubbleFish,crosswalk-project/chromium-crosswalk-efl,ltilve/chromium,littlstar/chromium.src,anirudhSK/chromium,krieger-od/nwjs_chromium.src,crosswalk-project/chromium-crosswalk-efl,Jonekee/chromium.src,TheTypoMaster/chromium-crosswalk,ChromiumWebApps/chromium,bright-sparks/chromium-spacewalk,mohamed--abdel-maksoud/chromium.src,timopulkkinen/BubbleFish,robclark/chromium,axinging/chromium-crosswalk,patrickm/chromium.src,Chilledheart/chromium,markYoungH/chromium.src,markYoungH/chromium.src,mogoweb/chromium-crosswalk,chuan9/chromium-crosswalk,hgl888/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,pozdnyakov/chromium-crosswalk,bright-sparks/chromium-spacewalk,zcbenz/cefode-chromium,dushu1203/chromium.src,Chilledheart/chromium,ChromiumWebApps/chromium,M4sse/chromium.src,axinging/chromium-crosswalk,hujiajie/pa-chromium,nacl-webkit/chrome_deps,patrickm/chromium.src,anirudhSK/chromium,bright-sparks/chromium-spacewalk,Just-D/chromium-1,littlstar/chromium.src,bright-sparks/chromium-spacewalk,ChromiumWebApps/chromium,M4sse/chromium.src,dednal/chromium.src,mohamed--abdel-maksoud/chromium.src,ltilve/chromium,keishi/chromium,dednal/chromium.src,keishi/chromium,ChromiumWebApps/chromium,rogerwang/chromium,M4sse/chromium.src,mogoweb/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,rogerwang/chromium,mohamed--abdel-maksoud/chromium.src,ondra-novak/chromium.src,rogerwang/chromium,rogerwang/chromium,robclark/chromium,jaruba/chromium.src,hgl888/chromium-crosswalk-efl,keishi/chromium
--- +++ @@ -23,3 +23,6 @@ results.extend(input_api.canned_checks.CheckChangeLintsClean( input_api, output_api, sources)) return results + +def GetPreferredTrySlaves(): + return ['win', 'linux', 'linux_view', 'linux_chromeos', 'mac']
a55f816072503241bd1ff4e953de12a7b48af4ac
backend/unimeet/helpers.py
backend/unimeet/helpers.py
from .models import School, User import re import string import random def get_school_list(): schools = School.objects.all() school_list = [] for s in schools: school = {} school['id'] = s.id school['name'] = s.name school['site'] = s.site school['university'] = s.university.name school['city'] = s.university.city.name school['country'] = s.university.city.country.name school_list.append(school) return school_list def get_school_by_email(email): for school in School.objects.all(): if re.match(school.mailRegex, email): return school return None def create_user(email, school): password = User.objects.make_random_password() token = ''.join(random.SystemRandom().choice(string.ascii_lowercase + string.digits) for _ in range(40)) user_obj = User.objects.create_user(email=email, school=school, password=password, token=token) user_obj.interestedInSchools.set(School.objects.all().values_list('id', flat=True)) user_obj.save() print 'Email:', email, 'Password:', password, 'Token:', token # TODO: Send signup mail to user
from .models import School, User import re import string import random from mail import send_mail def get_school_list(): schools = School.objects.all() school_list = [] for s in schools: school = {} school['id'] = s.id school['name'] = s.name school['site'] = s.site school['university'] = s.university.name school['city'] = s.university.city.name school['country'] = s.university.city.country.name school_list.append(school) return school_list def get_school_by_email(email): for school in School.objects.all(): if re.match(school.mailRegex, email): return school return None def create_user(email, school): password = User.objects.make_random_password() token = ''.join(random.SystemRandom().choice(string.ascii_lowercase + string.digits) for _ in range(40)) user_obj = User.objects.create_user(email=email, school=school, password=password, token=token) user_obj.interestedInSchools.set(School.objects.all().values_list('id', flat=True)) user_obj.save() send_mail(email, password, 'welcome')
Use send_mail in signup helper function
Use send_mail in signup helper function
Python
mit
dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet,dimkarakostas/unimeet
--- +++ @@ -2,6 +2,7 @@ import re import string import random +from mail import send_mail def get_school_list(): @@ -32,5 +33,4 @@ user_obj = User.objects.create_user(email=email, school=school, password=password, token=token) user_obj.interestedInSchools.set(School.objects.all().values_list('id', flat=True)) user_obj.save() - print 'Email:', email, 'Password:', password, 'Token:', token - # TODO: Send signup mail to user + send_mail(email, password, 'welcome')
4b1b5d0b71100fea17f127683a58533ef0e06fe9
bintools/splitter.py
bintools/splitter.py
import os # Splits a file using the dsplit mechanism def dsplit(fromfile, todir, chunksize = 1024): if not os.path.exists(todir): # caller handles errors os.mkdir(todir) # make dir, read/write parts original_file = os.path.basename(fromfile) filesize = os.path.getsize(fromfile) cont = True partnum = 0 while cont: if chunksize > filesize: cont = False chunksize = filesize chunk = __read_write_block(fromfile, chunksize) if not chunk: break partnum = partnum + 1 filename = os.path.join(todir, ('%s.part%d' % (original_file, partnum))) fileobj = open(filename, 'wb') fileobj.write(chunk) fileobj.close() chunksize *= 2 #### Private methods def __read_write_block(f, n): stream = open(f, 'rb') chunk = stream.read(n) stream.close() return chunk
import os # Splits a file using the dsplit mechanism def dsplit(fromfile, todir = os.getcwd(), offset = 0, limit = None, chunksize = 1024): if not os.path.exists(todir): # caller handles errors os.mkdir(todir) # make dir, read/write parts original_file = os.path.basename(fromfile) filesize = os.path.getsize(fromfile) cont = True partnum = 0 while cont: if chunksize > filesize: # Do 1 more read if chunksize > filesize cont = False chunksize = filesize partnum = partnum + 1 tofile = os.path.join(todir, ('%s.part%d' % (original_file, partnum))) chunk = __read_write_block(fromfile, chunksize, tofile) chunksize *= 2 #### Private methods def __read_write_block(fromfile, n, tofile, offset = 0): stream = open(fromfile, 'rb') chunk = stream.read(n) stream.close() if not chunk: return fileobj = open(tofile, 'wb') fileobj.write(chunk) fileobj.close() return fileobj
Refactor functions & add params
Refactor functions & add params
Python
apache-2.0
FernandoDoming/offset_finder
--- +++ @@ -1,7 +1,7 @@ import os # Splits a file using the dsplit mechanism -def dsplit(fromfile, todir, chunksize = 1024): +def dsplit(fromfile, todir = os.getcwd(), offset = 0, limit = None, chunksize = 1024): if not os.path.exists(todir): # caller handles errors os.mkdir(todir) # make dir, read/write parts @@ -11,21 +11,23 @@ partnum = 0 while cont: if chunksize > filesize: + # Do 1 more read if chunksize > filesize cont = False chunksize = filesize - chunk = __read_write_block(fromfile, chunksize) - if not chunk: break partnum = partnum + 1 - filename = os.path.join(todir, ('%s.part%d' % (original_file, partnum))) - fileobj = open(filename, 'wb') - fileobj.write(chunk) - fileobj.close() + tofile = os.path.join(todir, ('%s.part%d' % (original_file, partnum))) + chunk = __read_write_block(fromfile, chunksize, tofile) chunksize *= 2 #### Private methods -def __read_write_block(f, n): - stream = open(f, 'rb') +def __read_write_block(fromfile, n, tofile, offset = 0): + stream = open(fromfile, 'rb') chunk = stream.read(n) stream.close() - return chunk + if not chunk: return + + fileobj = open(tofile, 'wb') + fileobj.write(chunk) + fileobj.close() + return fileobj
fab9ff4d5d0f04f4ebfe86ed407b16ea73110a04
apps/package/templatetags/package_tags.py
apps/package/templatetags/package_tags.py
from datetime import timedelta from datetime import datetime from django import template from github2.client import Github from package.models import Package, Commit register = template.Library() github = Github() @register.filter def commits_over_52(package): current = datetime.now() weeks = [] commits = Commit.objects.filter(package=package).values_list('commit_date', flat=True) for week in range(52): weeks.append(len([x for x in commits if x < current and x > (current - timedelta(7))])) current -= timedelta(7) weeks.reverse() weeks = map(str, weeks) return ','.join(weeks)
from datetime import datetime, timedelta from django import template from package.models import Commit register = template.Library() @register.filter def commits_over_52(package): current = datetime.now() weeks = [] commits = Commit.objects.filter(package=package).values_list('commit_date', flat=True) for week in range(52): weeks.append(len([x for x in commits if x < current and x > (current - timedelta(7))])) current -= timedelta(7) weeks.reverse() weeks = map(str, weeks) return ','.join(weeks)
Clean up some imports in the package app's template_tags.py file.
Clean up some imports in the package app's template_tags.py file.
Python
mit
QLGu/djangopackages,pydanny/djangopackages,cartwheelweb/packaginator,nanuxbe/djangopackages,cartwheelweb/packaginator,audreyr/opencomparison,QLGu/djangopackages,nanuxbe/djangopackages,pydanny/djangopackages,miketheman/opencomparison,pydanny/djangopackages,QLGu/djangopackages,nanuxbe/djangopackages,miketheman/opencomparison,audreyr/opencomparison,benracine/opencomparison,cartwheelweb/packaginator,benracine/opencomparison
--- +++ @@ -1,15 +1,10 @@ -from datetime import timedelta -from datetime import datetime +from datetime import datetime, timedelta from django import template -from github2.client import Github - -from package.models import Package, Commit +from package.models import Commit register = template.Library() - -github = Github() @register.filter def commits_over_52(package):
d3a24fae87005b7f5c47657851b4341726494383
atest/resources/atest_variables.py
atest/resources/atest_variables.py
from os.path import abspath, dirname, join, normpath import locale import os import subprocess import robot __all__ = ['ROBOTPATH', 'ROBOT_VERSION', 'DATADIR', 'SYSTEM_ENCODING', 'CONSOLE_ENCODING'] ROBOTPATH = dirname(abspath(robot.__file__)) ROBOT_VERSION = robot.version.get_version() DATADIR = normpath(join(dirname(abspath(__file__)), '..', 'testdata')) SYSTEM_ENCODING = locale.getpreferredencoding(False) # Python 3.6+ uses UTF-8 internally on Windows. We want real console encoding. if os.name == 'nt': output = subprocess.check_output('chcp', shell=True, encoding='ASCII') CONSOLE_ENCODING = 'cp' + output.split()[-1] else: CONSOLE_ENCODING = locale.getdefaultlocale()[-1]
from os.path import abspath, dirname, join, normpath import locale import os import subprocess import robot __all__ = ['ROBOTPATH', 'ROBOT_VERSION', 'DATADIR', 'SYSTEM_ENCODING', 'CONSOLE_ENCODING'] ROBOTPATH = dirname(abspath(robot.__file__)) ROBOT_VERSION = robot.version.get_version() DATADIR = normpath(join(dirname(abspath(__file__)), '..', 'testdata')) SYSTEM_ENCODING = locale.getpreferredencoding(False) # Python 3.6+ uses UTF-8 internally on Windows. We want real console encoding. if os.name == 'nt': output = subprocess.check_output('chcp', shell=True, encoding='ASCII', errors='ignore') CONSOLE_ENCODING = 'cp' + output.split()[-1] else: CONSOLE_ENCODING = locale.getdefaultlocale()[-1]
Fix getting Windows system encoding on non-ASCII envs
atests: Fix getting Windows system encoding on non-ASCII envs
Python
apache-2.0
HelioGuilherme66/robotframework,robotframework/robotframework,HelioGuilherme66/robotframework,robotframework/robotframework,HelioGuilherme66/robotframework,robotframework/robotframework
--- +++ @@ -17,7 +17,8 @@ SYSTEM_ENCODING = locale.getpreferredencoding(False) # Python 3.6+ uses UTF-8 internally on Windows. We want real console encoding. if os.name == 'nt': - output = subprocess.check_output('chcp', shell=True, encoding='ASCII') + output = subprocess.check_output('chcp', shell=True, encoding='ASCII', + errors='ignore') CONSOLE_ENCODING = 'cp' + output.split()[-1] else: CONSOLE_ENCODING = locale.getdefaultlocale()[-1]
f21796d28ebf9328b68c6321d691b221457bafa6
jenkins/scripts/pypi-extract-universal.py
jenkins/scripts/pypi-extract-universal.py
#!/usr/bin/python # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # # Extract Python package name from setup.cfg import ConfigParser universal = False setup_cfg = ConfigParser.SafeConfigParser() setup_cfg.read("setup.cfg") if setup_cfg.has_section("wheel"): universal = setup_cfg.getboolean("wheel", "universal") if universal: print("py2.py3") else: print("py2")
#!/usr/bin/python # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. # # Extract Python package name from setup.cfg import ConfigParser universal = False setup_cfg = ConfigParser.SafeConfigParser() setup_cfg.read("setup.cfg") if setup_cfg.has_option("wheel", "universal"): universal = setup_cfg.getboolean("wheel", "universal") if universal: print("py2.py3") else: print("py2")
Fix universal extraction in non-universal wheel
Fix universal extraction in non-universal wheel The pypi-extract-universal.py would raise ConfigParser.NoOptionError when inspecting a setup.cfg with a wheel section but no universal option. Guard against this by actually testing whether the option is there rather than merely whether the section exists. Change-Id: I7a907464e5b1dc211c7e4c4c2f3e6375b7a6085d
Python
apache-2.0
Tesora/tesora-project-config,coolsvap/project-config,noorul/os-project-config,Tesora/tesora-project-config,dongwenjuan/project-config,dongwenjuan/project-config,openstack-infra/project-config,openstack-infra/project-config,anbangr/osci-project-config,coolsvap/project-config,anbangr/osci-project-config,noorul/os-project-config
--- +++ @@ -20,7 +20,7 @@ setup_cfg = ConfigParser.SafeConfigParser() setup_cfg.read("setup.cfg") -if setup_cfg.has_section("wheel"): +if setup_cfg.has_option("wheel", "universal"): universal = setup_cfg.getboolean("wheel", "universal") if universal: print("py2.py3")
757df7c04d862feb9067ae52c83875fc2e3aedf8
cla_backend/apps/cla_provider/admin/base.py
cla_backend/apps/cla_provider/admin/base.py
from django.contrib import admin from core.admin.modeladmin import OneToOneUserAdmin from ..models import Provider, ProviderAllocation, Staff, OutOfHoursRota from .forms import StaffAdminForm class StaffAdmin(OneToOneUserAdmin): model = Staff form = StaffAdminForm actions = None list_display = ( 'username_display', 'email_display', 'first_name_display', 'last_name_display', 'provider', 'is_active_display', 'is_manager' ) search_fields = [ 'user__username', 'user__first_name', 'user__last_name', 'user__email' ] class ProviderAllocationInline(admin.TabularInline): model = ProviderAllocation class ProviderAdmin(admin.ModelAdmin): actions = None inlines = [ProviderAllocationInline] fields = ( 'name', 'short_code', 'telephone_frontdoor', 'telephone_backdoor', 'email_address', 'active' ) list_display = ['name', 'law_categories', 'active'] def law_categories(self, obj): return u', '.join( obj.providerallocation_set.values_list('category__code', flat=True) ) admin.site.register(Provider, ProviderAdmin) admin.site.register(ProviderAllocation) admin.site.register(OutOfHoursRota) admin.site.register(Staff, StaffAdmin)
from django.contrib import admin from core.admin.modeladmin import OneToOneUserAdmin from ..models import Provider, ProviderAllocation, Staff, OutOfHoursRota from .forms import StaffAdminForm class StaffAdmin(OneToOneUserAdmin): model = Staff form = StaffAdminForm actions = None list_display = ( 'username_display', 'email_display', 'first_name_display', 'last_name_display', 'provider', 'is_active_display', 'is_manager' ) search_fields = [ 'user__username', 'user__first_name', 'user__last_name', 'user__email' ] class ProviderAllocationInline(admin.TabularInline): model = ProviderAllocation class ProviderAdmin(admin.ModelAdmin): actions = None inlines = [ProviderAllocationInline] fields = ( 'name', 'short_code', 'telephone_frontdoor', 'telephone_backdoor', 'email_address', 'active' ) list_display = ['name', 'law_categories', 'active'] def law_categories(self, obj): return u', '.join( obj.providerallocation_set.values_list('category__code', flat=True) ) admin.site.register(Provider, ProviderAdmin) admin.site.register(OutOfHoursRota) admin.site.register(Staff, StaffAdmin)
Disable ProviderAllocation admin page, still accessible from Provider Inlines
Disable ProviderAllocation admin page, still accessible from Provider Inlines
Python
mit
ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend
--- +++ @@ -44,6 +44,5 @@ admin.site.register(Provider, ProviderAdmin) -admin.site.register(ProviderAllocation) admin.site.register(OutOfHoursRota) admin.site.register(Staff, StaffAdmin)
f5d2cc8406b16fe2f0d4f640109dbb9a1ceba8d8
purchase_discount/models/purchase_order.py
purchase_discount/models/purchase_order.py
# -*- coding: utf-8 -*- # © 2004-2009 Tiny SPRL (<http://tiny.be>). # © 2015 Pedro M. Baeza # © 2016 ACSONE SA/NV (<http://acsone.eu>) # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from openerp import api, fields, models import openerp.addons.decimal_precision as dp class PurchaseOrderLine(models.Model): _inherit = "purchase.order.line" @api.depends('discount') def _compute_amount(self): prices = {} for line in self: if line.discount: prices[line.id] = line.price_unit line.price_unit *= (1 - line.discount / 100.0) super(PurchaseOrderLine, self)._compute_amount() # restore prices for line in self: if self.discount: line.price_unit = prices[line.id] discount = fields.Float( string='Discount (%)', digits_compute=dp.get_precision('Discount')) _sql_constraints = [ ('discount_limit', 'CHECK (discount <= 100.0)', 'Discount must be lower than 100%.'), ]
# -*- coding: utf-8 -*- # © 2004-2009 Tiny SPRL (<http://tiny.be>). # © 2015 Pedro M. Baeza # © 2016 ACSONE SA/NV (<http://acsone.eu>) # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from openerp import api, fields, models import openerp.addons.decimal_precision as dp class PurchaseOrderLine(models.Model): _inherit = "purchase.order.line" @api.depends('discount') def _compute_amount(self): prices = {} for line in self: if line.discount: prices[line.id] = line.price_unit line.price_unit *= (1 - line.discount / 100.0) super(PurchaseOrderLine, self)._compute_amount() # restore prices for line in self: if line.discount: line.price_unit = prices[line.id] discount = fields.Float( string='Discount (%)', digits_compute=dp.get_precision('Discount')) _sql_constraints = [ ('discount_limit', 'CHECK (discount <= 100.0)', 'Discount must be lower than 100%.'), ]
Fix issue with multiple lines
Fix issue with multiple lines
Python
agpl-3.0
OCA/purchase-workflow,OCA/purchase-workflow
--- +++ @@ -20,7 +20,7 @@ super(PurchaseOrderLine, self)._compute_amount() # restore prices for line in self: - if self.discount: + if line.discount: line.price_unit = prices[line.id] discount = fields.Float(
b8bf868d6ae7dbeb695dac36d5f72231d429d180
clone-vm.py
clone-vm.py
#!/usr/bin/env python import os import sys # Set arguments values if len(sys.argv) == 1: print "Usage: clone-vm.py [new-vm-name]" exit(1) else: new_vm_name = sys.argv[1] vms_home_dir = os.getenv('HOME') + '/Documents/Virtual\ Machines.localized/' vms_home_dir_not_escaped = os.getenv('HOME') + '/Documents/Virtual Machines.localized/' vm_source_vmx = vms_home_dir + 'base-centos-64.vmwarevm/base-centos-64.vmx' vm_dest_dir_not_escaped = vms_home_dir_not_escaped + new_vm_name + ".vmwarevm" vm_dest_dir = vms_home_dir + new_vm_name vm_dest_vmx = vm_dest_dir + '/' + new_vm_name + '.vmx' if not os.path.exists(vm_dest_dir_not_escaped): os.makedirs(vm_dest_dir_not_escaped) cmd = 'vmrun clone ' + vm_source_vmx + ' ' + vm_dest_vmx + ' linked -cloneName=' + new_vm_name print "[+] Creating new linked vm" os.system(cmd)
#!/usr/bin/env python import os, sys # Set arguments values if len(sys.argv) == 1: print "Usage: clone-vm.py [new-vm-name]" exit(1) else: new_vm_name = sys.argv[1] vms_path_dir = os.getenv('HOME') + '/Documents/Virtual Machines.localized/' vm_source_vmx = vms_home_dir + 'base-centos-64.vmwarevm/base-centos-64.vmx' vm_dest_dir = vms_home_dir + new_vm_name + ".vmwarevm" vm_dest_vmx = vm_dest_dir + '/' + new_vm_name + '.vmx' if not os.path.exists(vm_dest_dir): os.makedirs(vm_dest_dir) cmd = 'vmrun clone \'' + vm_source_vmx + '\' \'' + vm_dest_vmx + '\' linked -cloneName=' + new_vm_name print "[+] Creating new linked vm: " + new_vm_name os.system(cmd)
Delete escped variable, replaced with single quotes
Delete escped variable, replaced with single quotes
Python
apache-2.0
slariviere/py-vmrum,slariviere/py-vmrum
--- +++ @@ -1,7 +1,6 @@ #!/usr/bin/env python -import os -import sys +import os, sys # Set arguments values if len(sys.argv) == 1: @@ -10,19 +9,17 @@ else: new_vm_name = sys.argv[1] -vms_home_dir = os.getenv('HOME') + '/Documents/Virtual\ Machines.localized/' -vms_home_dir_not_escaped = os.getenv('HOME') + '/Documents/Virtual Machines.localized/' +vms_path_dir = os.getenv('HOME') + '/Documents/Virtual Machines.localized/' vm_source_vmx = vms_home_dir + 'base-centos-64.vmwarevm/base-centos-64.vmx' -vm_dest_dir_not_escaped = vms_home_dir_not_escaped + new_vm_name + ".vmwarevm" -vm_dest_dir = vms_home_dir + new_vm_name +vm_dest_dir = vms_home_dir + new_vm_name + ".vmwarevm" vm_dest_vmx = vm_dest_dir + '/' + new_vm_name + '.vmx' -if not os.path.exists(vm_dest_dir_not_escaped): - os.makedirs(vm_dest_dir_not_escaped) +if not os.path.exists(vm_dest_dir): + os.makedirs(vm_dest_dir) -cmd = 'vmrun clone ' + vm_source_vmx + ' ' + vm_dest_vmx + ' linked -cloneName=' + new_vm_name +cmd = 'vmrun clone \'' + vm_source_vmx + '\' \'' + vm_dest_vmx + '\' linked -cloneName=' + new_vm_name -print "[+] Creating new linked vm" +print "[+] Creating new linked vm: " + new_vm_name os.system(cmd)
f2a0c0c7329087421f6d3c237d2bb5f9633d180c
linear_math_tests/test_alignedobjectarray.py
linear_math_tests/test_alignedobjectarray.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ linear_math_tests.test_alignedobjectarray """ from __future__ import unicode_literals, print_function, absolute_import import unittest import math import bullet class ClassTestName(unittest.TestCase): def setUp(self): pass def tearDown(self): pass
#!/usr/bin/env python # -*- coding: utf-8 -*- """ linear_math_tests.test_alignedobjectarray """ from __future__ import unicode_literals, print_function, absolute_import import unittest import math import bullet class ClassTestName(unittest.TestCase): def setUp(self): self.a = bullet.btVector3Array() for i in range(10): self.a.append(bullet.btVector3(i, i+1, i+2)) self.b = bullet.btVector3Array() for i in range(10, 20): self.b.append(bullet.btVector3(i, i+1, i+2)) def test_assignment(self): self.a[0] = bullet.btVector3(21, 22, 23) self.assertEqual(self.a[0], bullet.btVector3(21, 22, 23)) def _slice(): self.a[0:3] = bullet.btVector3() self.assertRaises(RuntimeError, _slice) def tearDown(self): del self.a del self.b
Add some basic tests for assignment. Note that slicing is not supported
Add some basic tests for assignment. Note that slicing is not supported
Python
mit
Klumhru/boost-python-bullet,Klumhru/boost-python-bullet,Klumhru/boost-python-bullet
--- +++ @@ -13,7 +13,23 @@ class ClassTestName(unittest.TestCase): def setUp(self): - pass + self.a = bullet.btVector3Array() + for i in range(10): + self.a.append(bullet.btVector3(i, i+1, i+2)) + self.b = bullet.btVector3Array() + for i in range(10, 20): + self.b.append(bullet.btVector3(i, i+1, i+2)) + + def test_assignment(self): + self.a[0] = bullet.btVector3(21, 22, 23) + self.assertEqual(self.a[0], + bullet.btVector3(21, 22, 23)) + + def _slice(): + self.a[0:3] = bullet.btVector3() + + self.assertRaises(RuntimeError, _slice) def tearDown(self): - pass + del self.a + del self.b
941da2b1453cda2b981c5891fcc5d58c04df4544
eve_api/tasks.py
eve_api/tasks.py
from celery.decorators import task from eve_api.api_puller.accounts import import_eve_account from eve_api.app_defines import * from sso.tasks import update_user_access @task() def import_apikey(api_userid, api_key, user=None, force_cache=False): acc = import_eve_account(api_key, api_userid, force_cache=force_cache) donecorps = [] if acc and acc.api_status == API_STATUS_OK: if user and not acc.user: acc.user = user if acc.api_keytype == API_KEYTYPE_FULL and acc.characters.filter(director=1).count(): donecorps = [] for char in acc.characters.filter(director=1): if not char.corporation.id in donecorps: #pull_corp_members(acc.api_key, acc.api_user_id, char.id) char.corporation.query_and_update_corp() donecorps.append(char.corporation.id) for char in acc.characters.all(): try: if char.corporation.id not in donecorps: char.corporation.query_and_update_corp() donecorps.append(char.corporation.id) except: continue acc.save() if acc.user: update_user_access.delay(user=acc.user) return acc
import logging from celery.decorators import task from eve_api.api_puller.accounts import import_eve_account from eve_api.app_defines import * from sso.tasks import update_user_access @task() def import_apikey(api_userid, api_key, user=None, force_cache=False): l = logging.getLogger('import_apikey') l.info("Importing %s/%s" % (api_userid, api_key)) acc = import_eve_account(api_key, api_userid, force_cache=force_cache) donecorps = [] if acc and acc.api_status == API_STATUS_OK: if user and not acc.user: acc.user = user if acc.api_keytype == API_KEYTYPE_FULL and acc.characters.filter(director=1).count(): donecorps = [] for char in acc.characters.filter(director=1): if not char.corporation.id in donecorps: #pull_corp_members(acc.api_key, acc.api_user_id, char.id) char.corporation.query_and_update_corp() donecorps.append(char.corporation.id) for char in acc.characters.all(): try: if char.corporation.id not in donecorps: char.corporation.query_and_update_corp() donecorps.append(char.corporation.id) except: continue acc.save() if acc.user: update_user_access.delay(user=acc.user) return acc
Add some logging to the EVE API task
Add some logging to the EVE API task
Python
bsd-3-clause
nikdoof/test-auth
--- +++ @@ -1,3 +1,4 @@ +import logging from celery.decorators import task from eve_api.api_puller.accounts import import_eve_account from eve_api.app_defines import * @@ -5,6 +6,8 @@ @task() def import_apikey(api_userid, api_key, user=None, force_cache=False): + l = logging.getLogger('import_apikey') + l.info("Importing %s/%s" % (api_userid, api_key)) acc = import_eve_account(api_key, api_userid, force_cache=force_cache) donecorps = [] if acc and acc.api_status == API_STATUS_OK:
698a3fe81a15b40b95836426f9292365f9f57c9c
cartoframes/core/cartodataframe.py
cartoframes/core/cartodataframe.py
from geopandas import GeoDataFrame from ..utils.geom_utils import generate_index, generate_geometry class CartoDataFrame(GeoDataFrame): def __init__(self, *args, **kwargs): index_column = kwargs.pop('index_column', None) geom_column = kwargs.pop('geom_column', None) lnglat_column = kwargs.pop('lnglat_column', None) keep_index = kwargs.pop('keep_index', False) keep_geom = kwargs.pop('keep_geom', False) keep_lnglat = kwargs.pop('keep_lnglat', False) super(CartoDataFrame, self).__init__(*args, **kwargs) generate_index(self, index_column, keep_index) generate_geometry(self, geom_column, lnglat_column, keep_geom, keep_lnglat) @staticmethod def from_carto(*args, **kwargs): from ..io.carto import read_carto return read_carto(*args, **kwargs) def to_carto(*args, **kwargs): from ..io.carto import to_carto return to_carto(*args, **kwargs) def render(self, *args, **kwargs): from ..viz import Map, Layer return Map(Layer(self, *args, **kwargs))
from geopandas import GeoDataFrame from ..utils.geom_utils import generate_index, generate_geometry class CartoDataFrame(GeoDataFrame): def __init__(self, *args, **kwargs): index_column = kwargs.pop('index_column', None) geom_column = kwargs.pop('geom_column', None) lnglat_column = kwargs.pop('lnglat_column', None) keep_index = kwargs.pop('keep_index', False) keep_geom = kwargs.pop('keep_geom', False) keep_lnglat = kwargs.pop('keep_lnglat', False) super(CartoDataFrame, self).__init__(*args, **kwargs) generate_index(self, index_column, keep_index) generate_geometry(self, geom_column, lnglat_column, keep_geom, keep_lnglat) @staticmethod def from_carto(*args, **kwargs): from ..io.carto import read_carto return read_carto(*args, **kwargs) @classmethod def from_file(cls, filename, **kwargs): gdf = GeoDataFrame.from_file(filename, **kwargs) return cls(gdf) @classmethod def from_features(cls, features, **kwargs): gdf = GeoDataFrame.from_features(features, **kwargs) return cls(gdf) def to_carto(*args, **kwargs): from ..io.carto import to_carto return to_carto(*args, **kwargs) def render(self, *args, **kwargs): from ..viz import Map, Layer return Map(Layer(self, *args, **kwargs))
Add a wrapper for from_file/from_features methods
Add a wrapper for from_file/from_features methods
Python
bsd-3-clause
CartoDB/cartoframes,CartoDB/cartoframes
--- +++ @@ -23,6 +23,16 @@ from ..io.carto import read_carto return read_carto(*args, **kwargs) + @classmethod + def from_file(cls, filename, **kwargs): + gdf = GeoDataFrame.from_file(filename, **kwargs) + return cls(gdf) + + @classmethod + def from_features(cls, features, **kwargs): + gdf = GeoDataFrame.from_features(features, **kwargs) + return cls(gdf) + def to_carto(*args, **kwargs): from ..io.carto import to_carto return to_carto(*args, **kwargs)
9028bbd2f624196d6b28eaf7fcd3dccbcfca5f14
py/oldfart/handler.py
py/oldfart/handler.py
import http.server import os import oldfart.make __all__ = ['make_http_request_handler_class'] # The idea here is to modify the request handling by intercepting the # `send_head` call which is combines the common bits of GET and HEAD commands # and, more importantly, is the first method in the request handling process # that accesses the file system. Try to generate the resource to the file # system just before the parent, SimpleHTTPRequestHandle, needs it and we're # done. def _send_head(self): # FIXME: We die here if the directory doesn't exist ('make clean' # anyone?). All fixes seem ugly. Think about it. path = self.translate_path(self.path) target = os.path.relpath(path, self.maker.project_dir) if not os.path.isdir(path): retval, output = self.maker.make(target) if retval == oldfart.make.FAILURE: self.log_error('Building resource failed:\n%s', output) self.send_error(500, 'Could not generate resource') return None elif retval == oldfart.make.NO_RULE: self.log_message('No rule for building the resource') return super(self.__class__, self).send_head() def make_http_request_handler_class(name, maker): cls = type(name, (http.server.SimpleHTTPRequestHandler,), { 'maker': maker, 'send_head': _send_head }) return cls
import http.server import os import oldfart.make __all__ = ['make_http_request_handler_class'] # The idea here is to modify the request handling by intercepting the # `send_head` call which is combines the common bits of GET and HEAD commands # and, more importantly, is the first method in the request handling process # that accesses the file system. Try to generate the resource to the file # system just before the parent, SimpleHTTPRequestHandle, needs it and we're # done. def _send_head(self): # FIXME: We die here if the directory doesn't exist ('make clean' # anyone?). All fixes seem ugly. Think about it. path = self.translate_path(self.path) target = os.path.relpath(path, self.maker.project_dir) if not os.path.isdir(path): retval, output = self.maker.make(target) if retval == oldfart.make.FAILURE: self.log_error('Building resource "%s" failed:\n%s', self.path, output.rstrip()) self.send_error(500, 'Could not generate resource. See server log ' 'for details') return None elif retval == oldfart.make.NO_RULE: self.log_message('No rule for building the resource') return super(self.__class__, self).send_head() def make_http_request_handler_class(name, maker): cls = type(name, (http.server.SimpleHTTPRequestHandler,), { 'maker': maker, 'send_head': _send_head }) return cls
Improve logging of build failure
Improve logging of build failure
Python
bsd-3-clause
mjhanninen/oldfart,mjhanninen/oldfart,mjhanninen/oldfart
--- +++ @@ -22,8 +22,9 @@ if not os.path.isdir(path): retval, output = self.maker.make(target) if retval == oldfart.make.FAILURE: - self.log_error('Building resource failed:\n%s', output) - self.send_error(500, 'Could not generate resource') + self.log_error('Building resource "%s" failed:\n%s', self.path, output.rstrip()) + self.send_error(500, 'Could not generate resource. See server log ' + 'for details') return None elif retval == oldfart.make.NO_RULE: self.log_message('No rule for building the resource')
8f11aa020d5e539653120d5e895ea6f7b09392ce
cea/interfaces/dashboard/api/dashboard.py
cea/interfaces/dashboard/api/dashboard.py
from flask_restplus import Namespace, Resource, fields, abort import cea.config import cea.plots.cache api = Namespace('Dashboard', description='Dashboard plots') LAYOUTS = ['row', 'grid', 'map'] CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]} for c in cea.plots.categories.list_categories()} @api.route('/') class Dashboard(Resource): def get(self): """ Get Dashboards from yaml file """ config = cea.config.Configuration() plot_cache = cea.plots.cache.PlotCache(config) dashboards = cea.plots.read_dashboards(config, plot_cache) return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row', 'plots': [{'title': plot.title, 'scenario': plot.parameters['scenario-name']} for plot in d.plots]} for d in dashboards]
from flask_restplus import Namespace, Resource, fields, abort import cea.config import cea.plots.cache api = Namespace('Dashboard', description='Dashboard plots') LAYOUTS = ['row', 'grid', 'map'] CATEGORIES = {c.name: {'label': c.label, 'plots': [{'id': p.id(), 'name': p.name} for p in c.plots]} for c in cea.plots.categories.list_categories()} @api.route('/') class Dashboard(Resource): def get(self): """ Get Dashboards from yaml file """ config = cea.config.Configuration() plot_cache = cea.plots.cache.PlotCache(config) dashboards = cea.plots.read_dashboards(config, plot_cache) return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row', 'plots': [{'title': plot.title, 'scenario': plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None} for plot in d.plots]} for d in dashboards]
Allow 'scenario-name' to be null if it does not exist
Allow 'scenario-name' to be null if it does not exist
Python
mit
architecture-building-systems/CEAforArcGIS,architecture-building-systems/CEAforArcGIS
--- +++ @@ -22,4 +22,6 @@ dashboards = cea.plots.read_dashboards(config, plot_cache) return [{'name': d.name, 'description': d.description, 'layout': d.layout if d.layout in LAYOUTS else 'row', - 'plots': [{'title': plot.title, 'scenario': plot.parameters['scenario-name']} for plot in d.plots]} for d in dashboards] + 'plots': [{'title': plot.title, 'scenario': + plot.parameters['scenario-name'] if 'scenario-name' in plot.parameters.keys() else None} + for plot in d.plots]} for d in dashboards]
cf57bc6d564fdaf2af71f9eb8114b2487ae94867
meinberlin/config/settings/dev.py
meinberlin/config/settings/dev.py
from .base import * # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True for template_engine in TEMPLATES: template_engine['OPTIONS']['debug'] = True # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'qid$h1o8&wh#p(j)lifis*5-rf@lbiy8%^3l4x%@b$z(tli@ab' try: import debug_toolbar except ImportError: pass else: INSTALLED_APPS += ('debug_toolbar',) MIDDLEWARE += ('debug_toolbar.middleware.DebugToolbarMiddleware',) INTERNAL_IPS = ('127.0.0.1', 'localhost') try: from .local import * except ImportError: pass try: INSTALLED_APPS += tuple(ADDITIONAL_APPS) except NameError: pass
from .base import * # SECURITY WARNING: don't run with debug turned on in production! DEBUG = True for template_engine in TEMPLATES: template_engine['OPTIONS']['debug'] = True # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'qid$h1o8&wh#p(j)lifis*5-rf@lbiy8%^3l4x%@b$z(tli@ab' # FIXME: reenable after upgrade to wagtail 1.12 # see: https://github.com/jazzband/django-debug-toolbar/issues/950 # try: # import debug_toolbar # except ImportError: # pass # else: # INSTALLED_APPS += ('debug_toolbar',) # MIDDLEWARE += ('debug_toolbar.middleware.DebugToolbarMiddleware',) # # INTERNAL_IPS = ('127.0.0.1', 'localhost') try: from .local import * except ImportError: pass try: INSTALLED_APPS += tuple(ADDITIONAL_APPS) except NameError: pass
Disable django debug toolbar until wagtail 1.12 is released
Disable django debug toolbar until wagtail 1.12 is released see https://github.com/jazzband/django-debug-toolbar/issues/950 for reference
Python
agpl-3.0
liqd/a4-meinberlin,liqd/a4-meinberlin,liqd/a4-meinberlin,liqd/a4-meinberlin
--- +++ @@ -9,16 +9,17 @@ # SECURITY WARNING: keep the secret key used in production secret! SECRET_KEY = 'qid$h1o8&wh#p(j)lifis*5-rf@lbiy8%^3l4x%@b$z(tli@ab' - -try: - import debug_toolbar -except ImportError: - pass -else: - INSTALLED_APPS += ('debug_toolbar',) - MIDDLEWARE += ('debug_toolbar.middleware.DebugToolbarMiddleware',) - - INTERNAL_IPS = ('127.0.0.1', 'localhost') +# FIXME: reenable after upgrade to wagtail 1.12 +# see: https://github.com/jazzband/django-debug-toolbar/issues/950 +# try: +# import debug_toolbar +# except ImportError: +# pass +# else: +# INSTALLED_APPS += ('debug_toolbar',) +# MIDDLEWARE += ('debug_toolbar.middleware.DebugToolbarMiddleware',) +# +# INTERNAL_IPS = ('127.0.0.1', 'localhost') try: from .local import *
c79714249a0278b49a19a6a219328c4a74453c2d
cme/modules/MachineAccountQuota.py
cme/modules/MachineAccountQuota.py
from impacket.ldap import ldapasn1 as ldapasn1_impacket class CMEModule: ''' Module by Shutdown and Podalirius Initial module: https://github.com/ShutdownRepo/CrackMapExec-MachineAccountQuota Authors: Shutdown: @_nwodtuhs Podalirius: @podalirius_ ''' def options(self, context, module_options): pass name = 'MAQ' description = 'Retrieves the MachineAccountQuota domain-level attribute' supported_protocols = ['ldap'] opsec_safe = True multiple_hosts = False def on_login(self, context, connection): result = [] context.log.info('Getting the MachineAccountQuota') searchFilter = '(objectClass=*)' attributes = ['ms-DS-MachineAccountQuota'] result = connection.search(searchFilter, attributes, 1) for item in result: if isinstance(item, ldapasn1_impacket.SearchResultEntry) is not True: continue context.log.highlight("MachineAccountQuota: %d" % item['attributes'][0]['vals'][0])
from impacket.ldap import ldapasn1 as ldapasn1_impacket class CMEModule: ''' Module by Shutdown and Podalirius Initial module: https://github.com/ShutdownRepo/CrackMapExec-MachineAccountQuota Authors: Shutdown: @_nwodtuhs Podalirius: @podalirius_ ''' def options(self, context, module_options): pass name = 'MAQ' description = 'Retrieves the MachineAccountQuota domain-level attribute' supported_protocols = ['ldap'] opsec_safe = True multiple_hosts = False def on_login(self, context, connection): result = [] context.log.info('Getting the MachineAccountQuota') searchFilter = '(objectClass=*)' attributes = ['ms-DS-MachineAccountQuota'] result = connection.search(searchFilter, attributes) context.log.highlight("MachineAccountQuota: %d" % result[0]['attributes'][0]['vals'][0])
Remove error message when using MAQ module
Remove error message when using MAQ module
Python
bsd-2-clause
byt3bl33d3r/CrackMapExec
--- +++ @@ -22,13 +22,10 @@ opsec_safe = True multiple_hosts = False - def on_login(self, context, connection): + def on_login(self, context, connection): result = [] context.log.info('Getting the MachineAccountQuota') searchFilter = '(objectClass=*)' attributes = ['ms-DS-MachineAccountQuota'] - result = connection.search(searchFilter, attributes, 1) - for item in result: - if isinstance(item, ldapasn1_impacket.SearchResultEntry) is not True: - continue - context.log.highlight("MachineAccountQuota: %d" % item['attributes'][0]['vals'][0]) + result = connection.search(searchFilter, attributes) + context.log.highlight("MachineAccountQuota: %d" % result[0]['attributes'][0]['vals'][0])
7ed9b5dc23867381c34a77f31ccf4da5effbb0b0
tracpro/orgs_ext/migrations/0002_auto_20150724_1609.py
tracpro/orgs_ext/migrations/0002_auto_20150724_1609.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals import json from django.conf import settings from django.db import models, migrations def add_available_languages(apps, schema_editor): """Set default available_languages to all languages defined for this project.""" all_languages = [l[0] for l in settings.LANGUAGES] for org in apps.get_model('orgs', 'Org').objects.all(): updated = False config = json.loads(org.config) if not config.get('available_languages'): config['available_languages'] = all_languages org.config = json.dumps(config) updated = True if not org.default_language: org.default_language = settings.DEFAULT_LANGUAGE updated = True if updated: org.save() class Migration(migrations.Migration): dependencies = [ ('orgs_ext', '0001_initial'), ] operations = [ migrations.RunPython(add_available_languages, migrations.RunPython.noop), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals import json from django.conf import settings from django.db import models, migrations def add_available_languages(apps, schema_editor): """Set default available_languages to all languages defined for this project.""" all_languages = [l[0] for l in settings.LANGUAGES] for org in apps.get_model('orgs', 'Org').objects.all(): updated = False config = json.loads(org.config) if org.config else {} if not config.get('available_languages'): config['available_languages'] = all_languages org.config = json.dumps(config) updated = True if not org.default_language: org.default_language = settings.DEFAULT_LANGUAGE updated = True if updated: org.save() class Migration(migrations.Migration): dependencies = [ ('orgs_ext', '0001_initial'), ] operations = [ migrations.RunPython(add_available_languages, migrations.RunPython.noop), ]
Handle if org.config is None
Handle if org.config is None
Python
bsd-3-clause
xkmato/tracpro,rapidpro/tracpro,xkmato/tracpro,rapidpro/tracpro,xkmato/tracpro,xkmato/tracpro,rapidpro/tracpro
--- +++ @@ -12,7 +12,7 @@ all_languages = [l[0] for l in settings.LANGUAGES] for org in apps.get_model('orgs', 'Org').objects.all(): updated = False - config = json.loads(org.config) + config = json.loads(org.config) if org.config else {} if not config.get('available_languages'): config['available_languages'] = all_languages org.config = json.dumps(config)
505a88eaa461eb99d67b36692869b6d4025a054f
gapipy/models/address.py
gapipy/models/address.py
from .base import BaseModel class Address(BaseModel): _as_is_fields = ['city', 'latitude', 'longitude', 'postal_zip', 'street'] _resource_fields = [('country', 'Country')] def __repr__(self): return '<{0}: {1}, {2}>'.format( self.__class__.__name__, self.city, self.country.name)
from .base import BaseModel class Address(BaseModel): _as_is_fields = ['city', 'latitude', 'longitude', 'postal_zip', 'street'] _resource_fields = [ ('state', 'State'), ('country', 'Country') ] def __repr__(self): return '<{0}: {1}, {2}>'.format( self.__class__.__name__, self.city, self.country.name)
Add State to Address model definition.
Add State to Address model definition.
Python
mit
gadventures/gapipy
--- +++ @@ -3,7 +3,10 @@ class Address(BaseModel): _as_is_fields = ['city', 'latitude', 'longitude', 'postal_zip', 'street'] - _resource_fields = [('country', 'Country')] + _resource_fields = [ + ('state', 'State'), + ('country', 'Country') + ] def __repr__(self): return '<{0}: {1}, {2}>'.format(
481f4444e063d5559d396a5a26154b9ebde27248
formspree/app.py
formspree/app.py
import json import stripe import flask from flask import g from flask.ext.sqlalchemy import SQLAlchemy from flask.ext.login import LoginManager, current_user from flask_redis import Redis import settings DB = SQLAlchemy() redis_store = Redis() stripe.api_key = settings.STRIPE_SECRET_KEY import routes from users.models import User def configure_login(app): login_manager = LoginManager() login_manager.init_app(app) login_manager.login_view = 'register' @login_manager.user_loader def load_user(id): return User.query.get(int(id)) @app.before_request def before_request(): g.user = current_user def create_app(): app = flask.Flask(__name__) app.config.from_object(settings) DB.init_app(app) redis_store.init_app(app) routes.configure_routes(app) configure_login(app) app.jinja_env.filters['json'] = json.dumps return app
import json import stripe import os import flask from flask import g from flask.ext.sqlalchemy import SQLAlchemy from flask.ext.login import LoginManager, current_user from flask.ext.cdn import CDN from formspree import log from flask_redis import Redis import settings DB = SQLAlchemy() redis_store = Redis() stripe.api_key = settings.STRIPE_SECRET_KEY cdn = CDN() import routes from users.models import User def configure_login(app): login_manager = LoginManager() login_manager.init_app(app) login_manager.login_view = 'register' @login_manager.user_loader def load_user(id): return User.query.get(int(id)) @app.before_request def before_request(): g.user = current_user def create_app(): app = flask.Flask(__name__) app.config.from_object(settings) DB.init_app(app) redis_store.init_app(app) routes.configure_routes(app) configure_login(app) app.jinja_env.filters['json'] = json.dumps app.config['CDN_DOMAIN'] = os.getenv('CDN_URL') app.config['CDN_HTTPS'] = True cdn.init_app(app) return app
Allow for static contents served over CDN
Allow for static contents served over CDN
Python
agpl-3.0
asm-products/formspree,asm-products/formspree,asm-products/formspree,asm-products/formspree
--- +++ @@ -1,16 +1,20 @@ import json import stripe +import os import flask from flask import g from flask.ext.sqlalchemy import SQLAlchemy from flask.ext.login import LoginManager, current_user +from flask.ext.cdn import CDN +from formspree import log from flask_redis import Redis import settings DB = SQLAlchemy() redis_store = Redis() stripe.api_key = settings.STRIPE_SECRET_KEY +cdn = CDN() import routes from users.models import User @@ -39,5 +43,7 @@ configure_login(app) app.jinja_env.filters['json'] = json.dumps - + app.config['CDN_DOMAIN'] = os.getenv('CDN_URL') + app.config['CDN_HTTPS'] = True + cdn.init_app(app) return app
fbadf23356b40c36378cef8f3a9c8b382bce9e32
comics/core/admin.py
comics/core/admin.py
from django.contrib import admin from comics.core import models class ComicAdmin(admin.ModelAdmin): list_display = ('slug', 'name', 'language', 'url', 'rights') prepopulated_fields = { 'slug': ('name',) } class ReleaseAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'comic', 'pub_date', 'fetched') list_filter = ['pub_date', 'fetched', 'comic'] date_hierarchy = 'pub_date' exclude = ('images',) class ImageAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'file', 'height', 'width', 'fetched', 'title', 'text') list_filter = ['fetched', 'comic'] date_hierarchy = 'fetched' admin.site.register(models.Comic, ComicAdmin) admin.site.register(models.Release, ReleaseAdmin) admin.site.register(models.Image, ImageAdmin)
from django.contrib import admin from comics.core import models class ComicAdmin(admin.ModelAdmin): list_display = ('slug', 'name', 'language', 'url', 'rights', 'start_date', 'end_date', 'active') prepopulated_fields = { 'slug': ('name',) } class ReleaseAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'comic', 'pub_date', 'fetched') list_filter = ['pub_date', 'fetched', 'comic'] date_hierarchy = 'pub_date' exclude = ('images',) class ImageAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'file', 'height', 'width', 'fetched', 'title', 'text') list_filter = ['fetched', 'comic'] date_hierarchy = 'fetched' admin.site.register(models.Comic, ComicAdmin) admin.site.register(models.Release, ReleaseAdmin) admin.site.register(models.Image, ImageAdmin)
Include start date, end date, and active flag in comics list
Include start date, end date, and active flag in comics list
Python
agpl-3.0
jodal/comics,jodal/comics,jodal/comics,datagutten/comics,datagutten/comics,datagutten/comics,jodal/comics,datagutten/comics
--- +++ @@ -4,7 +4,8 @@ class ComicAdmin(admin.ModelAdmin): - list_display = ('slug', 'name', 'language', 'url', 'rights') + list_display = ('slug', 'name', 'language', 'url', 'rights', 'start_date', + 'end_date', 'active') prepopulated_fields = { 'slug': ('name',) }
f2f078a866c0185a6194b3ebc8b0e7090b8adeca
src/wirecloud/core/catalogue_manager.py
src/wirecloud/core/catalogue_manager.py
# -*- coding: utf-8 -*- # Copyright 2012 Universidad Politécnica de Madrid # This file is part of Wirecloud. # Wirecloud is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # Wirecloud is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with Wirecloud. If not, see <http://www.gnu.org/licenses/>. from catalogue.utils import add_resource_from_template from wirecloud.workspace.mashupTemplateGenerator import build_template_from_workspace from wirecloud.markets.utils import MarketManager from wirecloudcommons.utils.template import TemplateParser class WirecloudCatalogueManager(MarketManager): def __init__(self, options): pass def publish_mashup(self, endpoint, published_workspace, user, publish_options, request=None): template = TemplateParser(build_template_from_workspace(publish_options, published_workspace.workspace, user)) resource = add_resource_from_template(published_workspace.get_template_url(request), template, user) resource.publish = True resource.save()
# -*- coding: utf-8 -*- # Copyright 2012 Universidad Politécnica de Madrid # This file is part of Wirecloud. # Wirecloud is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # Wirecloud is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with Wirecloud. If not, see <http://www.gnu.org/licenses/>. from catalogue.utils import add_resource_from_template from wirecloud.workspace.mashupTemplateGenerator import build_template_from_workspace from wirecloud.markets.utils import MarketManager from wirecloudcommons.utils.template import TemplateParser class WirecloudCatalogueManager(MarketManager): def __init__(self, options): pass def publish_mashup(self, endpoint, published_workspace, user, publish_options, request=None): template = TemplateParser(build_template_from_workspace(publish_options, published_workspace.workspace, user)) resource = add_resource_from_template(published_workspace.get_template_url(request), template, user) resource.public = True resource.save()
Fix workspace publish on the local catalogue
Fix workspace publish on the local catalogue
Python
agpl-3.0
rockneurotiko/wirecloud,jpajuelo/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud,jpajuelo/wirecloud,rockneurotiko/wirecloud
--- +++ @@ -32,5 +32,5 @@ template = TemplateParser(build_template_from_workspace(publish_options, published_workspace.workspace, user)) resource = add_resource_from_template(published_workspace.get_template_url(request), template, user) - resource.publish = True + resource.public = True resource.save()
da729057ac482f4c03d4512a615ee86c9901bba9
glitch/config.py
glitch/config.py
# Basic config for stuff that can be easily changed, but which is git-managed. # See also apikeys_sample.py for the configs which are _not_ git-managed. #server_domain = "http://www.infiniteglitch.net" server_domain = "http://50.116.55.59" http_port = 8888 # Port for the main web site, in debug mode renderer_port = 81 # Port for the renderer (/all.mp3 and friends) # Track limits in seconds max_track_length = 400 min_track_length = 90
# Basic config for stuff that can be easily changed, but which is git-managed. # See also apikeys_sample.py for the configs which are _not_ git-managed. #server_domain = "http://www.infiniteglitch.net" server_domain = "http://50.116.55.59" http_port = 8888 # Port for the main web site, in debug mode renderer_port = 8889 # Port for the renderer (/all.mp3 and friends) # Track limits in seconds max_track_length = 400 min_track_length = 90
Revert to port 8889, as it appears that unicorn sets port 81.
Revert to port 8889, as it appears that unicorn sets port 81.
Python
artistic-2.0
MikeiLL/appension,MikeiLL/appension,MikeiLL/appension,MikeiLL/appension
--- +++ @@ -4,7 +4,7 @@ server_domain = "http://50.116.55.59" http_port = 8888 # Port for the main web site, in debug mode -renderer_port = 81 # Port for the renderer (/all.mp3 and friends) +renderer_port = 8889 # Port for the renderer (/all.mp3 and friends) # Track limits in seconds max_track_length = 400
80a28d495bc57c6866800d037cfc389050166319
tracpro/profiles/tests/factories.py
tracpro/profiles/tests/factories.py
import factory import factory.django import factory.fuzzy from tracpro.test.factory_utils import FuzzyEmail __all__ = ['User'] class User(factory.django.DjangoModelFactory): username = factory.fuzzy.FuzzyText() email = FuzzyEmail() class Meta: model = "auth.User" @factory.post_generation def password(self, create, extracted, **kwargs): password = extracted or "password" self.set_password(password) if create: self.save()
import factory import factory.django import factory.fuzzy from tracpro.test.factory_utils import FuzzyEmail __all__ = ['User'] class User(factory.django.DjangoModelFactory): username = factory.fuzzy.FuzzyText() email = FuzzyEmail() class Meta: model = "auth.User" @factory.post_generation def password(self, create, extracted, **kwargs): password = extracted or self.username self.set_password(password) if create: self.save()
Use username as password default
Use username as password default For compatibility with TracProTest.login()
Python
bsd-3-clause
xkmato/tracpro,rapidpro/tracpro,rapidpro/tracpro,xkmato/tracpro,xkmato/tracpro,rapidpro/tracpro,xkmato/tracpro
--- +++ @@ -17,7 +17,7 @@ @factory.post_generation def password(self, create, extracted, **kwargs): - password = extracted or "password" + password = extracted or self.username self.set_password(password) if create: self.save()
d43f87c18807853fde0a0e79828b5a8e7ab036fc
assess_isoform_quantification/set_isoform_frequencies.py
assess_isoform_quantification/set_isoform_frequencies.py
#!/usr/bin/python """Usage: set_isoform_frequencies [{help}] [{version}] {pro_file} {help_short} {help} Show this message. {version_short} {version} Show version. {pro_file} Flux simulator gene expression profile file. """ from docopt import docopt from options import validate_file_option from schema import SchemaError HELP_SHORT = "-h" HELP = "--help" VERSION_SHORT = "-v" VERSION = "--version" PRO_FILE = "<pro-file>" __doc__ = __doc__.format( help_short=HELP_SHORT, help=HELP, version_short=VERSION_SHORT, version=VERSION, pro_file=PRO_FILE) # Read in command-line options options = docopt(__doc__, version="set_isoform_frequencies v0.1") # Validate command-line options try: options[PRO_FILE] = validate_file_option( options[PRO_FILE], "Could not open expression profile file") except SchemaError as exc: exit(exc.code) print("Now do all the things.")
#!/usr/bin/python """Usage: set_isoform_frequencies [{help}] [{version}] {pro_file} {help_short} {help} Show this message. {version_short} {version} Show version. {pro_file} Flux simulator gene expression profile file. """ from docopt import docopt from options import validate_file_option from pandas import read_csv from schema import SchemaError HELP_SHORT = "-h" HELP = "--help" VERSION_SHORT = "-v" VERSION = "--version" PRO_FILE = "<pro-file>" LOCUS_COL = 'loc' TRANSCRIPT_ID_COL = 't_id' CODING_COL = 'c' LENGTH_COL = 'len' FRACTION_COL = 'f' NUM_TRANSCRIPTS_COL = 'n' UNKNOWN_COL_1 = 'u1' UNKNOWN_COL_2 = 'u2' PRO_FILE_COLUMNS = [ LOCUS_COL, TRANSCRIPT_ID_COL, CODING_COL, LENGTH_COL, FRACTION_COL, NUM_TRANSCRIPTS_COL, UNKNOWN_COL_1, UNKNOWN_COL_2] __doc__ = __doc__.format( help_short=HELP_SHORT, help=HELP, version_short=VERSION_SHORT, version=VERSION, pro_file=PRO_FILE) # Read in command-line options options = docopt(__doc__, version="set_isoform_frequencies v0.1") # Validate command-line options try: options[PRO_FILE] = validate_file_option( options[PRO_FILE], "Could not open expression profile file") except SchemaError as exc: exit(exc.code) df = read_csv(options[PRO_FILE], sep='\s*', names=PRO_FILE_COLUMNS, index_col=TRANSCRIPT_ID_COL)
Read in Flux Simulator expression profile with pandas.
Read in Flux Simulator expression profile with pandas.
Python
mit
lweasel/piquant,COMBINE-lab/piquant,lweasel/piquant
--- +++ @@ -10,6 +10,7 @@ from docopt import docopt from options import validate_file_option +from pandas import read_csv from schema import SchemaError HELP_SHORT = "-h" @@ -17,6 +18,25 @@ VERSION_SHORT = "-v" VERSION = "--version" PRO_FILE = "<pro-file>" + +LOCUS_COL = 'loc' +TRANSCRIPT_ID_COL = 't_id' +CODING_COL = 'c' +LENGTH_COL = 'len' +FRACTION_COL = 'f' +NUM_TRANSCRIPTS_COL = 'n' +UNKNOWN_COL_1 = 'u1' +UNKNOWN_COL_2 = 'u2' + +PRO_FILE_COLUMNS = [ + LOCUS_COL, + TRANSCRIPT_ID_COL, + CODING_COL, + LENGTH_COL, + FRACTION_COL, + NUM_TRANSCRIPTS_COL, + UNKNOWN_COL_1, + UNKNOWN_COL_2] __doc__ = __doc__.format( help_short=HELP_SHORT, @@ -36,4 +56,5 @@ except SchemaError as exc: exit(exc.code) -print("Now do all the things.") +df = read_csv(options[PRO_FILE], sep='\s*', + names=PRO_FILE_COLUMNS, index_col=TRANSCRIPT_ID_COL)
156c049cc3965f969ee252dc5859cf0713bcbe27
grip/__init__.py
grip/__init__.py
"""\ Grip ---- Render local readme files before sending off to Github. :copyright: (c) 2012 by Joe Esposito. :license: MIT, see LICENSE for more details. """ __version__ = '1.2.0' from . import command from .server import default_filenames, serve from .renderer import render_content, render_page
"""\ Grip ---- Render local readme files before sending off to Github. :copyright: (c) 2012 by Joe Esposito. :license: MIT, see LICENSE for more details. """ __version__ = '1.2.0' from . import command from .renderer import render_content, render_page from .server import default_filenames, create_app, serve from .exporter import export
Add create_app and export to API.
Add create_app and export to API.
Python
mit
jbarreras/grip,jbarreras/grip,joeyespo/grip,mgoddard-pivotal/grip,ssundarraj/grip,ssundarraj/grip,mgoddard-pivotal/grip,joeyespo/grip
--- +++ @@ -12,5 +12,6 @@ from . import command -from .server import default_filenames, serve from .renderer import render_content, render_page +from .server import default_filenames, create_app, serve +from .exporter import export
d49d7dc7943e25be8497a7cbed059aa04fe76e7d
nefi2_main/nefi2/model/algorithms/_alg.py
nefi2_main/nefi2/model/algorithms/_alg.py
# -*- coding: utf-8 -*- """ This class represents an interface of an image processing algorithm. The class abstracts algorithm interface from user so he can fully focus on his algorithm implementation. """ __author__ = "p.shkadzko@gmail.com" class Algorithm: def __init__(self): """ Algorithm class Instance vars: self.modified -- True if Algorithm settings were modified self.belongs -- A step name to which current algorithm belongs """ self.modified = False # for debugging only print '> Algorithm: I am "%s" algorithm' % self.name def belongs(self): """Return a step name to which current algorithm belongs.""" return self.parent def process(self, image): """ A user must override this method in order to comply with the interface. """ raise NotImplementedError def get_name(self): return self.name def set_modified(self): """Set True if method settings were modified.""" print '> Algorithm: "%s" was modified.' % (self.name) self.modified = True def get_modified(self): return self.modified def report_pip(self): pass if __name__ == '__main__': pass
# -*- coding: utf-8 -*- """ This class represents an interface of an image processing algorithm. The class abstracts algorithm interface from user so he can fully focus on his algorithm implementation. """ __author__ = "p.shkadzko@gmail.com" class Algorithm: def __init__(self): """ Algorithm class Instance vars: self.modified -- True if Algorithm settings were modified self.belongs -- A step name to which current algorithm belongs """ self.modified = True # for debugging only print '> Algorithm: I am "%s" algorithm' % self.name def belongs(self): """Return a step name to which current algorithm belongs.""" return self.parent def process(self, image): """ A user must override this method in order to comply with the interface. """ raise NotImplementedError def get_name(self): return self.name def set_modified(self): """Set True if method settings were modified.""" print '> Algorithm: "%s" was modified.' % (self.name) self.modified = True def get_modified(self): return self.modified def report_pip(self): pass if __name__ == '__main__': pass
Change modfied variable to true
Change modfied variable to true
Python
bsd-2-clause
LumPenPacK/NetworkExtractionFromImages,LumPenPacK/NetworkExtractionFromImages,LumPenPacK/NetworkExtractionFromImages,LumPenPacK/NetworkExtractionFromImages,LumPenPacK/NetworkExtractionFromImages
--- +++ @@ -17,7 +17,7 @@ self.modified -- True if Algorithm settings were modified self.belongs -- A step name to which current algorithm belongs """ - self.modified = False + self.modified = True # for debugging only print '> Algorithm: I am "%s" algorithm' % self.name
0ddad675169952f861f164d7ce311c83dccd51e0
invoice/admin.py
invoice/admin.py
from django.contrib import admin from django.conf.urls.defaults import patterns from invoice.models import Invoice, InvoiceItem from invoice.views import pdf_view from invoice.forms import InvoiceAdminForm class InvoiceItemInline(admin.TabularInline): model = InvoiceItem class InvoiceAdmin(admin.ModelAdmin): inlines = [InvoiceItemInline,] fieldsets = ( (None, { 'fields': ('user', 'address', 'status') }), ('Dates', { 'fields': ('invoice_date', 'due_date', 'paid_date') }), ) search_fields = ('invoice_id', 'user__username') list_filter = ('status',) list_display = ( 'invoice_id', 'user', 'invoice_date', 'due_date', 'status', 'total_amount', ) form = InvoiceAdminForm def get_urls(self): urls = super(InvoiceAdmin, self).get_urls() return patterns('', (r'^(.+)/pdf/$', self.admin_site.admin_view(pdf_view)) ) + urls admin.site.register(Invoice, InvoiceAdmin)
from django.contrib import admin from django.conf.urls.defaults import patterns from invoice.models import Invoice, InvoiceItem from invoice.views import pdf_view from invoice.forms import InvoiceAdminForm class InvoiceItemInline(admin.TabularInline): model = InvoiceItem class InvoiceAdmin(admin.ModelAdmin): inlines = [InvoiceItemInline,] fieldsets = ( (None, { 'fields': ('user', 'address', 'status') }), ('Dates', { 'fields': ('invoice_date', 'due_date', 'paid_date') }), ) search_fields = ('invoice_id', 'user__username') list_filter = ('status',) list_display = ( 'invoice_id', 'user', 'invoice_date', 'due_date', 'paid_date', 'status', 'total_amount', ) form = InvoiceAdminForm def get_urls(self): urls = super(InvoiceAdmin, self).get_urls() return patterns('', (r'^(.+)/pdf/$', self.admin_site.admin_view(pdf_view)) ) + urls admin.site.register(Invoice, InvoiceAdmin)
Add paid_date to invoice list.
Add paid_date to invoice list.
Python
bsd-3-clause
simonluijk/django-invoice,Chris7/django-invoice,Chris7/django-invoice
--- +++ @@ -26,6 +26,7 @@ 'user', 'invoice_date', 'due_date', + 'paid_date', 'status', 'total_amount', )
09c2e6fff38e5c47391c0f8e948089e3efd26337
serfnode/handler/file_utils.py
serfnode/handler/file_utils.py
import os from tempfile import mkstemp import time class atomic_write(object): """Perform an atomic write to a file. Use as:: with atomic_write('/my_file') as f: f.write('foo') """ def __init__(self, filepath): """ :type filepath: str """ self.filepath = filepath def __enter__(self): """ :rtype: File """ _, self.temp = mkstemp(dir=os.getcwd()) self.f = open(self.temp, 'w') return self.f def __exit__(self, exc_type, exc_val, exc_tb): self.f.close() if exc_type is None: os.rename(self.temp, self.filepath) def wait_for_file(filepath, sleep_interval=0.1): """Wait for the existence of a file. Warning: use ``atomic_write`` to write the file, since this function doesn't check that the file is complete. :type filepath: str :type sleep_interval: float :rtype: None """ while not os.path.exists(filepath): time.sleep(sleep_interval)
import os from tempfile import mkstemp import time class atomic_write(object): """Perform an atomic write to a file. Use as:: with atomic_write('/my_file') as f: f.write('foo') """ def __init__(self, filepath): """ :type filepath: str """ self.filepath = filepath def __enter__(self): """ :rtype: File """ _, self.temp = mkstemp(dir=os.getcwd()) self.f = open(self.temp, 'w') return self.f def __exit__(self, exc_type, exc_val, exc_tb): self.f.close() if exc_type is None: os.rename(self.temp, self.filepath) def wait_for_files(*filepath, sleep_interval=0.1): """Wait for the existence of files. Warning: use ``atomic_write`` to write the file, since this function doesn't check that the file is complete. :type filepath: str :type sleep_interval: float :rtype: None """ while not all(os.path.exists(f) for f in filepath): time.sleep(sleep_interval)
Allow waiting for multiple files
Allow waiting for multiple files
Python
mit
waltermoreira/serfnode,waltermoreira/serfnode,waltermoreira/serfnode
--- +++ @@ -33,8 +33,8 @@ os.rename(self.temp, self.filepath) -def wait_for_file(filepath, sleep_interval=0.1): - """Wait for the existence of a file. +def wait_for_files(*filepath, sleep_interval=0.1): + """Wait for the existence of files. Warning: use ``atomic_write`` to write the file, since this function doesn't check that the file is complete. @@ -43,5 +43,5 @@ :type sleep_interval: float :rtype: None """ - while not os.path.exists(filepath): + while not all(os.path.exists(f) for f in filepath): time.sleep(sleep_interval)
eb61e5c989cda3f5e021150f91561a88ba6db73e
setuptools/tests/py26compat.py
setuptools/tests/py26compat.py
import sys import unittest import tarfile try: # provide skipIf for Python 2.4-2.6 skipIf = unittest.skipIf except AttributeError: def skipIf(condition, reason): def skipper(func): def skip(*args, **kwargs): return if condition: return skip return func return skipper def _tarfile_open_ex(*args, **kwargs): """ Extend result as a context manager. """ res = tarfile.open(*args, **kwargs) res.__exit__ = lambda exc_type, exc_value, traceback: res.close() res.__enter__ = lambda: res return res tarfile_open = _tarfile_open_ex if sys.version_info < (2,7) else tarfile.open
import sys import unittest import tarfile import contextlib try: # provide skipIf for Python 2.4-2.6 skipIf = unittest.skipIf except AttributeError: def skipIf(condition, reason): def skipper(func): def skip(*args, **kwargs): return if condition: return skip return func return skipper def _tarfile_open_ex(*args, **kwargs): """ Extend result as a context manager. """ return contextlib.closing(tarfile.open(*args, **kwargs)) tarfile_open = _tarfile_open_ex if sys.version_info < (2,7) else tarfile.open
Use contextlib.closing on tarfile compat shim
Use contextlib.closing on tarfile compat shim
Python
mit
pypa/setuptools,pypa/setuptools,pypa/setuptools
--- +++ @@ -1,6 +1,7 @@ import sys import unittest import tarfile +import contextlib try: # provide skipIf for Python 2.4-2.6 @@ -19,9 +20,6 @@ """ Extend result as a context manager. """ - res = tarfile.open(*args, **kwargs) - res.__exit__ = lambda exc_type, exc_value, traceback: res.close() - res.__enter__ = lambda: res - return res + return contextlib.closing(tarfile.open(*args, **kwargs)) tarfile_open = _tarfile_open_ex if sys.version_info < (2,7) else tarfile.open
29b3ea81fad91b9cdba150a74aeeb43b40dc0d67
numba/cuda/tests/cudadrv/test_profiler.py
numba/cuda/tests/cudadrv/test_profiler.py
import unittest from numba.cuda.testing import ContextResettingTestCase from numba import cuda from numba.cuda.testing import skip_on_cudasim @skip_on_cudasim('CUDA Profiler unsupported in the simulator') class TestProfiler(ContextResettingTestCase): def test_profiling(self): with cuda.profiling(): a = cuda.device_array(10) del a with cuda.profiling(): a = cuda.device_array(100) del a if __name__ == '__main__': unittest.main()
import unittest from numba.cuda.testing import ContextResettingTestCase from numba import cuda from numba.cuda.testing import skip_on_cudasim, xfail_with_cuda_python @skip_on_cudasim('CUDA Profiler unsupported in the simulator') @xfail_with_cuda_python class TestProfiler(ContextResettingTestCase): def test_profiling(self): with cuda.profiling(): a = cuda.device_array(10) del a with cuda.profiling(): a = cuda.device_array(100) del a if __name__ == '__main__': unittest.main()
Revert "Re-enable profiler with CUDA Python"
Revert "Re-enable profiler with CUDA Python" This reverts commit 0a7a8d891b946ad7328ac83854b18935f3cf23d6. The Conda packages for the NVIDIA bindings do not support this API - only the Github source version supports it, the profiler tests must be disabled.
Python
bsd-2-clause
IntelLabs/numba,numba/numba,seibert/numba,IntelLabs/numba,cpcloud/numba,cpcloud/numba,IntelLabs/numba,numba/numba,cpcloud/numba,numba/numba,seibert/numba,seibert/numba,cpcloud/numba,numba/numba,IntelLabs/numba,seibert/numba,cpcloud/numba,IntelLabs/numba,seibert/numba,numba/numba
--- +++ @@ -1,10 +1,11 @@ import unittest from numba.cuda.testing import ContextResettingTestCase from numba import cuda -from numba.cuda.testing import skip_on_cudasim +from numba.cuda.testing import skip_on_cudasim, xfail_with_cuda_python @skip_on_cudasim('CUDA Profiler unsupported in the simulator') +@xfail_with_cuda_python class TestProfiler(ContextResettingTestCase): def test_profiling(self): with cuda.profiling():
4c6c7067c05026b0e40020b0be58de83aa79e4f5
entrypoint.py
entrypoint.py
import bcrypt import sys if len(sys.argv) < 2: print('Error: please provide a password.', file=sys.stderr) sys.exit(2) password = sys.argv[1] strength = None if len(sys.argv) > 2: strength = int(sys.argv[2]) if strength: salt = bcrypt.gensalt(rounds=strength) else: salt = bcrypt.gensalt() password = password.encode('utf-8') bhash = bcrypt.hashpw(password, salt) print(bhash.decode('utf-8'))
import bcrypt import sys if len(sys.argv) < 2: print('Error: please provide a password.', file=sys.stderr) sys.exit(2) password = sys.argv[1] strength = None if len(sys.argv) > 2: strength = int(sys.argv[2]) if strength: salt = bcrypt.gensalt(rounds=strength) else: salt = bcrypt.gensalt() password = password.encode('utf-8') bhash = bcrypt.hashpw(password, salt) print(bhash.decode('utf-8'), end='', flush=True) print(file=sys.stderr)
Remove newline character from stdout.
Remove newline character from stdout.
Python
apache-2.0
JohnStarich/docker-bcrypt
--- +++ @@ -14,5 +14,6 @@ salt = bcrypt.gensalt() password = password.encode('utf-8') bhash = bcrypt.hashpw(password, salt) -print(bhash.decode('utf-8')) +print(bhash.decode('utf-8'), end='', flush=True) +print(file=sys.stderr)
f10d2b0aa6da7347435ad1036c0e53e67c89d362
dougrain.py
dougrain.py
#!/usr/bin/python import urlparse import curie import link class Document(object): def expand_curie(self, link): return self.curie.expand(link) @classmethod def from_object(cls, o, relative_to_url=None, parent_curie=None): if isinstance(o, list): return map(lambda x: cls.from_object(x, relative_to_url), o) result = Document() result.attrs = o result.__dict__.update(o) result.links = {} for key, value in o.get("_links", {}).iteritems(): result.links[key] = link.Link.from_object(value, relative_to_url) if 'self' in result.links: result.url = result.links['self'].url result.curie = curie.CurieCollection(relative_to_url) if parent_curie is not None: result.curie.update(parent_curie) curies = result.links.get('curie', []) if not isinstance(curies, list): curies = [curies] for curie_dict in curies: result.curie[curie_dict.name] = curie_dict.href result.embedded = {} for key, value in o.get("_embedded", {}).iteritems(): result.embedded[key] = cls.from_object(value, relative_to_url, result.curie) return result
#!/usr/bin/python import urlparse import curie import link class Document(object): def __init__(self, o, relative_to_url, parent_curie=None): self.attrs = o self.__dict__.update(o) self.links = {} for key, value in o.get("_links", {}).iteritems(): self.links[key] = link.Link.from_object(value, relative_to_url) if 'self' in self.links: self.url = self.links['self'].url self.curie = curie.CurieCollection(relative_to_url) if parent_curie is not None: self.curie.update(parent_curie) curies = self.links.get('curie', []) if not isinstance(curies, list): curies = [curies] for curie_dict in curies: self.curie[curie_dict.name] = curie_dict.href self.embedded = {} for key, value in o.get("_embedded", {}).iteritems(): self.embedded[key] = self.__class__.from_object(value, relative_to_url, self.curie) def expand_curie(self, link): return self.curie.expand(link) @classmethod def from_object(cls, o, relative_to_url=None, parent_curie=None): if isinstance(o, list): return map(lambda x: cls.from_object(x, relative_to_url), o) return cls(o, relative_to_url, parent_curie)
Move some construction to __init__.
Refactor: Move some construction to __init__.
Python
mit
wharris/dougrain
--- +++ @@ -4,6 +4,33 @@ import link class Document(object): + def __init__(self, o, relative_to_url, parent_curie=None): + self.attrs = o + self.__dict__.update(o) + self.links = {} + + for key, value in o.get("_links", {}).iteritems(): + self.links[key] = link.Link.from_object(value, relative_to_url) + + if 'self' in self.links: + self.url = self.links['self'].url + + self.curie = curie.CurieCollection(relative_to_url) + if parent_curie is not None: + self.curie.update(parent_curie) + + curies = self.links.get('curie', []) + if not isinstance(curies, list): + curies = [curies] + for curie_dict in curies: + self.curie[curie_dict.name] = curie_dict.href + + self.embedded = {} + for key, value in o.get("_embedded", {}).iteritems(): + self.embedded[key] = self.__class__.from_object(value, + relative_to_url, + self.curie) + def expand_curie(self, link): return self.curie.expand(link) @@ -13,32 +40,7 @@ if isinstance(o, list): return map(lambda x: cls.from_object(x, relative_to_url), o) - result = Document() - result.attrs = o - result.__dict__.update(o) - result.links = {} + return cls(o, relative_to_url, parent_curie) - for key, value in o.get("_links", {}).iteritems(): - result.links[key] = link.Link.from_object(value, relative_to_url) - if 'self' in result.links: - result.url = result.links['self'].url - result.curie = curie.CurieCollection(relative_to_url) - if parent_curie is not None: - result.curie.update(parent_curie) - - curies = result.links.get('curie', []) - if not isinstance(curies, list): - curies = [curies] - for curie_dict in curies: - result.curie[curie_dict.name] = curie_dict.href - - result.embedded = {} - for key, value in o.get("_embedded", {}).iteritems(): - result.embedded[key] = cls.from_object(value, - relative_to_url, - result.curie) - - return result -
410c47921da205c1628cdff771f3385546edd503
src/engine/SCons/Platform/darwin.py
src/engine/SCons/Platform/darwin.py
"""engine.SCons.Platform.darwin Platform-specific initialization for Mac OS X systems. There normally shouldn't be any need to import this module directly. It will usually be imported through the generic SCons.Platform.Platform() selection method. """ # # Copyright (c) 2001, 2002, 2003, 2004 Steven Knight # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be included # in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # import posix import os def generate(env): posix.generate(env) env['SHLIBSUFFIX'] = '.dylib'
"""engine.SCons.Platform.darwin Platform-specific initialization for Mac OS X systems. There normally shouldn't be any need to import this module directly. It will usually be imported through the generic SCons.Platform.Platform() selection method. """ # # __COPYRIGHT__ # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the # "Software"), to deal in the Software without restriction, including # without limitation the rights to use, copy, modify, merge, publish, # distribute, sublicense, and/or sell copies of the Software, and to # permit persons to whom the Software is furnished to do so, subject to # the following conditions: # # The above copyright notice and this permission notice shall be included # in all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY # KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE # WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # __revision__ = "__FILE__ __REVISION__ __DATE__ __DEVELOPER__" import posix import os def generate(env): posix.generate(env) env['SHLIBSUFFIX'] = '.dylib'
Fix __COPYRIGHT__ and __REVISION__ in new Darwin module.
Fix __COPYRIGHT__ and __REVISION__ in new Darwin module.
Python
mit
Distrotech/scons,Distrotech/scons,Distrotech/scons,Distrotech/scons,Distrotech/scons
--- +++ @@ -8,7 +8,7 @@ """ # -# Copyright (c) 2001, 2002, 2003, 2004 Steven Knight +# __COPYRIGHT__ # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and associated documentation files (the @@ -30,6 +30,8 @@ # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. # +__revision__ = "__FILE__ __REVISION__ __DATE__ __DEVELOPER__" + import posix import os
f664b85c3bc68612d19ae0fc762ca314f6517b00
src/constants.py
src/constants.py
#!/usr/bin/env python TRAJECTORY = 'linear' CONTROLLER = 'euler' # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 60.0 MAX_V = 0.075 MAX_W = 1.25 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 MAX_V = 0.11 MAX_W = 1.25 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 MAX_V = 0.055 MAX_W = 1.20 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) RESULTS_DIRECTORY = '../txt_results/'
#!/usr/bin/env python TRAJECTORY = 'linear' CONTROLLER = 'euler' # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 SIMULATION_TIME_IN_SECONDS = 0.0 if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 60.0 MAX_V = 0.075 MAX_W = 1.25 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 MAX_V = 0.11 MAX_W = 1.25 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 MAX_V = 0.055 MAX_W = 1.20 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) RESULTS_DIRECTORY = '../txt_results/'
Add default value for simulation time
Add default value for simulation time
Python
mit
bit0001/trajectory_tracking,bit0001/trajectory_tracking
--- +++ @@ -16,7 +16,7 @@ K_I_W = 1.25 K_D_W = 0.000 - +SIMULATION_TIME_IN_SECONDS = 0.0 if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 60.0 MAX_V = 0.075