commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
82e964dab398caee75c3174f86593ab6cfa7dbaf
src/constants.py
src/constants.py
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 40.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 80.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
Increase simulation time for linear trajectory
Increase simulation time for linear trajectory
Python
mit
bit0001/trajectory_tracking,bit0001/trajectory_tracking
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 40.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 Increase simulation time for linear trajectory
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 80.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
<commit_before>#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 40.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 <commit_msg>Increase simulation time for linear trajectory<commit_after>
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 80.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 40.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 Increase simulation time for linear trajectory#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 80.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
<commit_before>#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 40.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000 <commit_msg>Increase simulation time for linear trajectory<commit_after>#!/usr/bin/env python TRAJECTORY = 'linear' if TRAJECTORY == 'linear': SIMULATION_TIME_IN_SECONDS = 80.0 elif TRAJECTORY == 'circular': SIMULATION_TIME_IN_SECONDS = 120.0 elif TRAJECTORY == 'squared': SIMULATION_TIME_IN_SECONDS = 160.0 DELTA_T = 0.1 # this is the sampling time STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T) # control constants K_X = 0.90 K_Y = 0.90 K_THETA = 0.90 # PID control constants K_P_V = 0.2 K_I_V = 1.905 K_D_V = 0.00 K_P_W = 0.45 K_I_W = 1.25 K_D_W = 0.000
63e61d9123a4abeb85d83545da351ed0abf792aa
src/core/urls.py
src/core/urls.py
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ]
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] try: if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] except AttributeError: pass
Handle installs not using new settings engine
Handle installs not using new settings engine
Python
agpl-3.0
BirkbeckCTP/janeway,BirkbeckCTP/janeway,BirkbeckCTP/janeway,BirkbeckCTP/janeway
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] Handle installs not using new settings engine
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] try: if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] except AttributeError: pass
<commit_before>__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] <commit_msg>Handle installs not using new settings engine<commit_after>
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] try: if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] except AttributeError: pass
__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] Handle installs not using new settings engine__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] try: if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] except AttributeError: pass
<commit_before>__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] <commit_msg>Handle installs not using new settings engine<commit_after>__copyright__ = "Copyright 2017 Birkbeck, University of London" __author__ = "Martin Paul Eve & Andy Byers" __license__ = "AGPL v3" __maintainer__ = "Birkbeck Centre for Technology and Publishing" from django.conf.urls import include, url from django.contrib import admin from django.views.generic import TemplateView from django.conf import settings from django.views.static import serve from press import views as press_views include('events.registration') urlpatterns = [ url(r'^$', press_views.index, name='website_index'), url(r'^admin/', include(admin.site.urls)), url(r'^summernote/', include('django_summernote.urls')), url(r'', include('core.include_urls')), ] try: if settings.DEBUG or settings.IN_TEST_RUNNER: import debug_toolbar urlpatterns += [ url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}), url(r'^404/$', TemplateView.as_view(template_name='404.html')), url(r'^500/$', TemplateView.as_view(template_name='500.html')), url(r'^__debug__/', include(debug_toolbar.urls)), url(r'^hijack/', include('hijack.urls', namespace='hijack')), ] except AttributeError: pass
1a5c1ea0815d30048d7dbce56adf2503b9c82c28
moksha/widgets/container/tests/test_container.py
moksha/widgets/container/tests/test_container.py
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert '<div id="test" ' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert 'Moksha Container' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
Update our container test case
Update our container test case
Python
apache-2.0
ralphbean/moksha,pombredanne/moksha,pombredanne/moksha,lmacken/moksha,mokshaproject/moksha,mokshaproject/moksha,lmacken/moksha,mokshaproject/moksha,pombredanne/moksha,mokshaproject/moksha,ralphbean/moksha,ralphbean/moksha,lmacken/moksha,pombredanne/moksha
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert '<div id="test" ' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered Update our container test case
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert 'Moksha Container' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
<commit_before># This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert '<div id="test" ' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered <commit_msg>Update our container test case<commit_after>
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert 'Moksha Container' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert '<div id="test" ' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered Update our container test case# This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert 'Moksha Container' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
<commit_before># This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert '<div id="test" ' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered <commit_msg>Update our container test case<commit_after># This file is part of Moksha. # Copyright (C) 2008-2009 Red Hat, Inc. # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. from tw.api import Widget from moksha.widgets.container import MokshaContainer class TestContainer: def setUp(self): self.w = MokshaContainer('test') def test_render_widget(self): assert 'Moksha Container' in self.w() def test_widget_content(self): """ Ensure we can render a container with another widget """ class MyWidget(Widget): template = """ Hello World! """ assert 'Hello World!' in self.w(content=MyWidget('mywidget')) def test_container_classes(self): rendered = self.w(**dict(skin3=True, stikynote=True, draggable=True, resizable=True)) assert 'class="containerPlus draggable resizable"' in rendered, rendered
35ffe6bb97a30970d4bc3c265b6337712669ee09
githubsetupircnotifications.py
githubsetupircnotifications.py
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: print('Failed to sign into github') sys.exit(1) org = github.organization(args.organization) if org is None: print('Organization "{}" does not appear to exist'.format(args.org)) sys.exit(1) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def error(message): print(message) sys.exit(1) def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: error('Failed to sign into github') org = github.organization(args.organization) if org is None: error('Organization "{}" does not appear to exist'.format(args.org)) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
Add error function to remove duplicate code
Add error function to remove duplicate code
Python
mit
kragniz/github-setup-irc-notifications
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: print('Failed to sign into github') sys.exit(1) org = github.organization(args.organization) if org is None: print('Organization "{}" does not appear to exist'.format(args.org)) sys.exit(1) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events) Add error function to remove duplicate code
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def error(message): print(message) sys.exit(1) def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: error('Failed to sign into github') org = github.organization(args.organization) if org is None: error('Organization "{}" does not appear to exist'.format(args.org)) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
<commit_before>""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: print('Failed to sign into github') sys.exit(1) org = github.organization(args.organization) if org is None: print('Organization "{}" does not appear to exist'.format(args.org)) sys.exit(1) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events) <commit_msg>Add error function to remove duplicate code<commit_after>
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def error(message): print(message) sys.exit(1) def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: error('Failed to sign into github') org = github.organization(args.organization) if org is None: error('Organization "{}" does not appear to exist'.format(args.org)) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: print('Failed to sign into github') sys.exit(1) org = github.organization(args.organization) if org is None: print('Organization "{}" does not appear to exist'.format(args.org)) sys.exit(1) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events) Add error function to remove duplicate code""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def error(message): print(message) sys.exit(1) def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: error('Failed to sign into github') org = github.organization(args.organization) if org is None: error('Organization "{}" does not appear to exist'.format(args.org)) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
<commit_before>""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: print('Failed to sign into github') sys.exit(1) org = github.organization(args.organization) if org is None: print('Organization "{}" does not appear to exist'.format(args.org)) sys.exit(1) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events) <commit_msg>Add error function to remove duplicate code<commit_after>""" github-setup-irc-notifications - Configure all repositories in an organization with irc notifications """ import argparse import getpass import sys import github3 def error(message): print(message) sys.exit(1) def main(): parser = argparse.ArgumentParser() parser.add_argument('--username') parser.add_argument('--password') parser.add_argument('organization') parser.add_argument('channel') args = parser.parse_args() if args.password is None: password = getpass.getpass( 'Password for github user "{}":'.format(args.username)) else: password = args.password github = github3.login(args.username, password=password) if github is None: error('Failed to sign into github') org = github.organization(args.organization) if org is None: error('Organization "{}" does not appear to exist'.format(args.org)) conf = {'nickserv_password': '', 'no_colors': '0', 'password': '', 'branch_regexes': '', 'room': args.channel, 'ssl': '0', 'port': '', 'branches': '', 'server': 'chat.freenode.net', 'long_url': '0', 'notice': '0', 'message_without_join': '1', 'nick': 'github' } events = [ 'push', 'delete', 'create', 'issues', 'pull_request' ] for r in org.iter_repos(): r.create_hook('irc', conf, events=events)
859a23790968c84cdbc4fa7467957a3a1ed1e069
greatbigcrane/project/forms.py
greatbigcrane/project/forms.py
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): name = forms.CharField(initial="django") settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
Add a name for the recipe section
Add a name for the recipe section
Python
apache-2.0
pnomolos/greatbigcrane,pnomolos/greatbigcrane
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm } Add a name for the recipe section
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): name = forms.CharField(initial="django") settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
<commit_before>""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm } <commit_msg>Add a name for the recipe section<commit_after>
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): name = forms.CharField(initial="django") settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm } Add a name for the recipe section""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): name = forms.CharField(initial="django") settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
<commit_before>""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm } <commit_msg>Add a name for the recipe section<commit_after>""" Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """ from django import forms from project.models import Project class ProjectForm(forms.ModelForm): class Meta: model = Project class DjangoRecipeForm(forms.Form): name = forms.CharField(initial="django") settings = forms.CharField() version = forms.ChoiceField(choices=[ ("trunk", "trunk"), ("1.2", "1.2"), ("1.1.2", "1.1.2"), ("1.1", "1.1"), ("1.0.4", "1.04"), ("0.96", "0.96"), ]) eggs = forms.CharField() project = forms.CharField() extra_paths = forms.CharField() fcgi = forms.BooleanField() wsgi = forms.BooleanField() recipe_form_map = { 'djangorecipe': DjangoRecipeForm }
61cf5a4ab4d7b9e0cb95925acc633aa7cb156d59
taggit/models.py
taggit/models.py
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField() def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField(unique=True, max_length=100) def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
Make slug unique and fix the max_length.
Make slug unique and fix the max_length.
Python
bsd-3-clause
izquierdo/django-taggit,theatlantic/django-taggit2,doselect/django-taggit,gem/django-taggit,theatlantic/django-taggit,Maplecroft/django-taggit,benjaminrigaud/django-taggit,kminkov/django-taggit,twig/django-taggit,cimani/django-taggit,decibyte/django-taggit,IRI-Research/django-taggit,eugena/django-taggit,theatlantic/django-taggit2,laanlabs/django-taggit,kaedroho/django-taggit,orbitvu/django-taggit,7kfpun/django-taggit,adrian-sgn/django-taggit,guoqiao/django-taggit,tamarmot/django-taggit,vhf/django-taggit,nealtodd/django-taggit,decibyte/django-taggit,theatlantic/django-taggit,Eksmo/django-taggit
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField() def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag) Make slug unique and fix the max_length.
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField(unique=True, max_length=100) def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
<commit_before>from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField() def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag) <commit_msg>Make slug unique and fix the max_length.<commit_after>
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField(unique=True, max_length=100) def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField() def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag) Make slug unique and fix the max_length.from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField(unique=True, max_length=100) def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
<commit_before>from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField() def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag) <commit_msg>Make slug unique and fix the max_length.<commit_after>from django.contrib.contenttypes.models import ContentType from django.contrib.contenttypes.generic import GenericForeignKey from django.db import models from django.template.defaultfilters import slugify class Tag(models.Model): name = models.CharField(max_length=100) slug = models.SlugField(unique=True, max_length=100) def __unicode__(self): return self.name def save(self, *args, **kwargs): if not self.pk: self.slug = slugify(self.name) super(Tag, self).save(*args, **kwargs) class TaggedItem(models.Model): object_id = models.IntegerField() content_type = models.ForeignKey(ContentType) content_object = GenericForeignKey() tag = models.ForeignKey(Tag, related_name="items") def __unicode__(self): return "%s tagged with %s" % (self.content_object, self.tag)
9afc0f35b718e11418c22e2f60e07f7a9ee9aaa3
core/commands/log_graph.py
core/commands/log_graph.py
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
Add `--all` to graph command to show other branches (esp. origin)
Add `--all` to graph command to show other branches (esp. origin)
Python
mit
asfaltboy/GitSavvy,asfaltboy/GitSavvy,jmanuel1/GitSavvy,divmain/GitSavvy,ddevlin/GitSavvy,theiviaxx/GitSavvy,divmain/GitSavvy,stoivo/GitSavvy,dvcrn/GitSavvy,theiviaxx/GitSavvy,ralic/GitSavvy,ddevlin/GitSavvy,ralic/GitSavvy,divmain/GitSavvy,stoivo/GitSavvy,dreki/GitSavvy,stoivo/GitSavvy,asfaltboy/GitSavvy,ddevlin/GitSavvy,dvcrn/GitSavvy,jmanuel1/GitSavvy,dreki/GitSavvy
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph}) Add `--all` to graph command to show other branches (esp. origin)
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
<commit_before>from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph}) <commit_msg>Add `--all` to graph command to show other branches (esp. origin)<commit_after>
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph}) Add `--all` to graph command to show other branches (esp. origin)from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
<commit_before>from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph}) <commit_msg>Add `--all` to graph command to show other branches (esp. origin)<commit_after>from sublime_plugin import WindowCommand, TextCommand from ..git_command import GitCommand LOG_GRAPH_TITLE = "GRAPH" class GsLogGraphCommand(WindowCommand, GitCommand): """ Open a new window displaying an ASCII-graphic representation of the repo's branch relationships. """ def run(self): repo_path = self.repo_path view = self.window.new_file() view.settings().set("git_savvy.log_graph_view", True) view.settings().set("git_savvy.repo_path", repo_path) view.set_name(LOG_GRAPH_TITLE) view.set_scratch(True) view.set_read_only(True) view.run_command("gs_log_graph_initialize") class GsLogGraphInitializeCommand(TextCommand, GitCommand): def run(self, edit): branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate") self.view.run_command("gs_replace_view_text", {"text": branch_graph})
7db62eefb0601098db6eff9becf566f51fee9be4
pythonforandroid/recipes/setuptools/__init__.py
pythonforandroid/recipes/setuptools/__init__.py
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
Use zip instead of tar.gz to obtain setuptools
Use zip instead of tar.gz to obtain setuptools Newer releases of setuptools on pypi are only available as zip archives. In order to support higher versions of setuptools, zip should be preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0 will fail.
Python
mit
kivy/python-for-android,rnixx/python-for-android,germn/python-for-android,rnixx/python-for-android,PKRoma/python-for-android,rnixx/python-for-android,rnixx/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,germn/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,germn/python-for-android,rnixx/python-for-android,germn/python-for-android,kronenpj/python-for-android,kivy/python-for-android,germn/python-for-android,rnixx/python-for-android,kivy/python-for-android,kivy/python-for-android,PKRoma/python-for-android,germn/python-for-android,kivy/python-for-android,kronenpj/python-for-android
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe() Use zip instead of tar.gz to obtain setuptools Newer releases of setuptools on pypi are only available as zip archives. In order to support higher versions of setuptools, zip should be preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0 will fail.
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
<commit_before>from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe() <commit_msg>Use zip instead of tar.gz to obtain setuptools Newer releases of setuptools on pypi are only available as zip archives. In order to support higher versions of setuptools, zip should be preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0 will fail.<commit_after>
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe() Use zip instead of tar.gz to obtain setuptools Newer releases of setuptools on pypi are only available as zip archives. In order to support higher versions of setuptools, zip should be preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0 will fail.from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
<commit_before>from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe() <commit_msg>Use zip instead of tar.gz to obtain setuptools Newer releases of setuptools on pypi are only available as zip archives. In order to support higher versions of setuptools, zip should be preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0 will fail.<commit_after>from pythonforandroid.recipe import PythonRecipe class SetuptoolsRecipe(PythonRecipe): version = '18.3.1' url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip' depends = [('python2', 'python3crystax')] call_hostpython_via_targetpython = False install_in_hostpython = True recipe = SetuptoolsRecipe()
b3757884bdaa6e488d54ee51f943dbb3578ea469
stores/forms.py
stores/forms.py
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( max_length=4, widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")}) ) state = forms.ChoiceField(choices=STATE_CHOICES)
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( widget=forms.TextInput( attrs={'placeholder': _("Enter your postcode or suburb...")} ) ) state = forms.ChoiceField(choices=STATE_CHOICES)
Remove limit for search field
Remove limit for search field
Python
bsd-3-clause
django-oscar/django-oscar-stores,django-oscar/django-oscar-stores,django-oscar/django-oscar-stores
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( max_length=4, widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")}) ) state = forms.ChoiceField(choices=STATE_CHOICES)Remove limit for search field
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( widget=forms.TextInput( attrs={'placeholder': _("Enter your postcode or suburb...")} ) ) state = forms.ChoiceField(choices=STATE_CHOICES)
<commit_before>from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( max_length=4, widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")}) ) state = forms.ChoiceField(choices=STATE_CHOICES)<commit_msg>Remove limit for search field<commit_after>
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( widget=forms.TextInput( attrs={'placeholder': _("Enter your postcode or suburb...")} ) ) state = forms.ChoiceField(choices=STATE_CHOICES)
from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( max_length=4, widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")}) ) state = forms.ChoiceField(choices=STATE_CHOICES)Remove limit for search fieldfrom django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( widget=forms.TextInput( attrs={'placeholder': _("Enter your postcode or suburb...")} ) ) state = forms.ChoiceField(choices=STATE_CHOICES)
<commit_before>from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( max_length=4, widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")}) ) state = forms.ChoiceField(choices=STATE_CHOICES)<commit_msg>Remove limit for search field<commit_after>from django import forms from django.db.models import get_model from django.utils.translation import ugettext as _ StoreAddress = get_model('stores', 'StoreAddress') class StoreSearchForm(forms.Form): STATE_CHOICES = ( (_('VIC'), _('Victoria')), (_('NSW'), _('New South Wales')), (_('SA'), _('South Australia')), (_('TAS'), _('Tasmania')), (_('QLD'), _('Queensland')), (_('NT'), _('Northern Territory')), ) location = forms.CharField(widget=forms.HiddenInput) store_search = forms.CharField( widget=forms.TextInput( attrs={'placeholder': _("Enter your postcode or suburb...")} ) ) state = forms.ChoiceField(choices=STATE_CHOICES)
fe50886a42bf7fa5e3217134e1f7a732960ab2d9
nbgrader/tests/apps/test_nbgrader_generate_config.py
nbgrader/tests/apps/test_nbgrader_generate_config.py
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") with open("nbgrader_config.py") as f: contents = f.read() # This was missing in issue #1089 assert "AssignLatePenalties" in contents # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
Add assertion for issue gh-1089
Add assertion for issue gh-1089
Python
bsd-3-clause
jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jhamrick/nbgrader,jhamrick/nbgrader
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1) Add assertion for issue gh-1089
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") with open("nbgrader_config.py") as f: contents = f.read() # This was missing in issue #1089 assert "AssignLatePenalties" in contents # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
<commit_before>import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1) <commit_msg>Add assertion for issue gh-1089<commit_after>
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") with open("nbgrader_config.py") as f: contents = f.read() # This was missing in issue #1089 assert "AssignLatePenalties" in contents # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1) Add assertion for issue gh-1089import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") with open("nbgrader_config.py") as f: contents = f.read() # This was missing in issue #1089 assert "AssignLatePenalties" in contents # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
<commit_before>import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1) <commit_msg>Add assertion for issue gh-1089<commit_after>import os from .. import run_nbgrader from .base import BaseTestApp class TestNbGraderGenerateConfig(BaseTestApp): def test_help(self): """Does the help display without error?""" run_nbgrader(["generate_config", "--help-all"]) def test_generate_config(self): """Is the config file properly generated?""" # it already exists, because we create it in conftest.py os.remove("nbgrader_config.py") # try recreating it run_nbgrader(["generate_config"]) assert os.path.isfile("nbgrader_config.py") with open("nbgrader_config.py") as f: contents = f.read() # This was missing in issue #1089 assert "AssignLatePenalties" in contents # does it fail if it already exists? run_nbgrader(["generate_config"], retcode=1)
5d673ed6cf8f810ee22bc701f6927f9580b4ec34
rest-api/config.py
rest-api/config.py
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' STAGING_CLIENT_ID = '101582376895035372811' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
Add the staging client id too.
Add the staging client id too.
Python
bsd-3-clause
all-of-us/raw-data-repository,all-of-us/raw-data-repository,all-of-us/raw-data-repository
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ] Add the staging client id too.
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' STAGING_CLIENT_ID = '101582376895035372811' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
<commit_before>"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ] <commit_msg>Add the staging client id too.<commit_after>
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' STAGING_CLIENT_ID = '101582376895035372811' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ] Add the staging client id too."""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' STAGING_CLIENT_ID = '101582376895035372811' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
<commit_before>"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ] <commit_msg>Add the staging client id too.<commit_after>"""Configuration parameters. Contains things such as the database to connect to. """ CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr' CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE CLOUDSQL_USER = 'api' PYTHON_TEST_CLIENT_ID = '116540421226121250670' STAGING_CLIENT_ID = '101582376895035372811' ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID] # TODO: Move all authentication into the datastore. ALLOWED_USERS = [ 'test-client@pmi-rdr-api-test.iam.gserviceaccount.com', 'pmi-hpo-staging@appspot.gserviceaccount.com', ]
d8a861c47df6b41c27f2ec43474766284ba728af
bot/logger/message_sender/reusable/limiter/group.py
bot/logger/message_sender/reusable/limiter/group.py
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued()
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() def notify_about_to_send_message(self): for limiter in self.limiters: limiter.notify_about_to_send_message()
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
Python
agpl-3.0
alvarogzp/telegram-bot,alvarogzp/telegram-bot
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() def notify_about_to_send_message(self): for limiter in self.limiters: limiter.notify_about_to_send_message()
<commit_before>from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() <commit_msg>Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters<commit_after>
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() def notify_about_to_send_message(self): for limiter in self.limiters: limiter.notify_about_to_send_message()
from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limitersfrom bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() def notify_about_to_send_message(self): for limiter in self.limiters: limiter.notify_about_to_send_message()
<commit_before>from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() <commit_msg>Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters<commit_after>from bot.logger.message_sender.message_builder import MessageBuilder from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter class ReusableMessageLimiterGroup(ReusableMessageLimiter): def __init__(self, *limiters: ReusableMessageLimiter): self.limiters = limiters def should_issue_new_message_pre_add(self, new_text): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text)) def should_issue_new_message_post_add(self, builder: MessageBuilder): return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder)) def __any_limiter(self, func: callable): return any((func(limiter) for limiter in self.limiters)) def notify_new_message_issued(self): for limiter in self.limiters: limiter.notify_new_message_issued() def notify_about_to_send_message(self): for limiter in self.limiters: limiter.notify_about_to_send_message()
769fa2c0c777ac88702e6b3802de4909c8f8df22
sh_app/forms.py
sh_app/forms.py
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state', 'is_private') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
Remove isPrivate field on league form
Remove isPrivate field on league form
Python
mit
skill-huddle/skill-huddle,skill-huddle/skill-huddle
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state', 'is_private') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, } Remove isPrivate field on league form
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
<commit_before>from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state', 'is_private') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, } <commit_msg>Remove isPrivate field on league form<commit_after>
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state', 'is_private') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, } Remove isPrivate field on league formfrom django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
<commit_before>from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state', 'is_private') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, } <commit_msg>Remove isPrivate field on league form<commit_after>from django import forms from django.forms import Textarea from sh_app.models import User, SH_User, League, Suggestion class UserForm(forms.ModelForm): password = forms.CharField(widget=forms.PasswordInput()) class Meta: model = User fields = ('username', 'email', 'password') class SH_UserForm(forms.ModelForm): class Meta: model = SH_User fields = ('first_name', 'last_name') class LeagueForm(forms.ModelForm): class Meta: model = League fields = ('name', 'description', 'country', 'city', 'state') help_texts = { 'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length) } widgets = { 'description': Textarea } class SuggestionForm(forms.ModelForm): class Meta: model = Suggestion fields = ('name', 'description', 'voting_ends') help_texts = { 'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length) } widgets = { 'description': Textarea, }
6dd52ba31141f28e1f37e32f8c3de6932ed49b4f
make_mozilla/base/tests/assertions.py
make_mozilla/base/tests/assertions.py
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function, name = '', kwargs = {}): resolved_route = resolve(url) ok_(resolved_route.func is view_function) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function_or_class, name = '', kwargs = {}): resolved_route = resolve(url) ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class)) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
Allow assert_routing to cope with Feed instances used as route endpoints
Allow assert_routing to cope with Feed instances used as route endpoints
Python
bsd-3-clause
mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function, name = '', kwargs = {}): resolved_route = resolve(url) ok_(resolved_route.func is view_function) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url) Allow assert_routing to cope with Feed instances used as route endpoints
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function_or_class, name = '', kwargs = {}): resolved_route = resolve(url) ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class)) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
<commit_before>from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function, name = '', kwargs = {}): resolved_route = resolve(url) ok_(resolved_route.func is view_function) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url) <commit_msg>Allow assert_routing to cope with Feed instances used as route endpoints<commit_after>
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function_or_class, name = '', kwargs = {}): resolved_route = resolve(url) ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class)) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function, name = '', kwargs = {}): resolved_route = resolve(url) ok_(resolved_route.func is view_function) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url) Allow assert_routing to cope with Feed instances used as route endpointsfrom nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function_or_class, name = '', kwargs = {}): resolved_route = resolve(url) ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class)) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
<commit_before>from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function, name = '', kwargs = {}): resolved_route = resolve(url) ok_(resolved_route.func is view_function) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url) <commit_msg>Allow assert_routing to cope with Feed instances used as route endpoints<commit_after>from nose.tools import eq_, ok_ from django.core.urlresolvers import resolve, reverse def assert_routing(url, view_function_or_class, name = '', kwargs = {}): resolved_route = resolve(url) ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class)) if kwargs: eq_(resolved_route.kwargs, kwargs) if name: eq_(reverse(name, kwargs = kwargs), url) def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False): status_codes = {True: 301, False: 302} expected_redirect_url = reverse(name, kwargs = kwargs) eq_(response.status_code, status_codes[permanent]) eq_(response['Location'], expected_redirect_url)
8f530fac6e16dcae18aae8696b5a30117137f9f3
rest/messages/sms-conversation-tracking/example-1.py
rest/messages/sms-conversation-tracking/example-1.py
from flask import Flask, request, redirect, session import twilio.twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" counter = session.get('counter', 0) # increment the counter counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" message = "".join([name, " has messaged ", request.values.get('To'), " ", str(counter), " times."]) resp = twilio.twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
from flask import Flask, request, session from twilio import twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" # Increment the counter counter = session.get('counter', 0) counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" # Build our reply message = '{} has messaged {} {} times.' \ .format(name, request.values.get('To'), counter) # Put it in a TwiML response resp = twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
Clean up Python example for sms-conversation-tracking
Clean up Python example for sms-conversation-tracking
Python
mit
teoreteetik/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets
from flask import Flask, request, redirect, session import twilio.twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" counter = session.get('counter', 0) # increment the counter counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" message = "".join([name, " has messaged ", request.values.get('To'), " ", str(counter), " times."]) resp = twilio.twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)Clean up Python example for sms-conversation-tracking
from flask import Flask, request, session from twilio import twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" # Increment the counter counter = session.get('counter', 0) counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" # Build our reply message = '{} has messaged {} {} times.' \ .format(name, request.values.get('To'), counter) # Put it in a TwiML response resp = twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
<commit_before>from flask import Flask, request, redirect, session import twilio.twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" counter = session.get('counter', 0) # increment the counter counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" message = "".join([name, " has messaged ", request.values.get('To'), " ", str(counter), " times."]) resp = twilio.twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)<commit_msg>Clean up Python example for sms-conversation-tracking<commit_after>
from flask import Flask, request, session from twilio import twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" # Increment the counter counter = session.get('counter', 0) counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" # Build our reply message = '{} has messaged {} {} times.' \ .format(name, request.values.get('To'), counter) # Put it in a TwiML response resp = twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
from flask import Flask, request, redirect, session import twilio.twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" counter = session.get('counter', 0) # increment the counter counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" message = "".join([name, " has messaged ", request.values.get('To'), " ", str(counter), " times."]) resp = twilio.twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)Clean up Python example for sms-conversation-trackingfrom flask import Flask, request, session from twilio import twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" # Increment the counter counter = session.get('counter', 0) counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" # Build our reply message = '{} has messaged {} {} times.' \ .format(name, request.values.get('To'), counter) # Put it in a TwiML response resp = twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
<commit_before>from flask import Flask, request, redirect, session import twilio.twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" counter = session.get('counter', 0) # increment the counter counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" message = "".join([name, " has messaged ", request.values.get('To'), " ", str(counter), " times."]) resp = twilio.twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)<commit_msg>Clean up Python example for sms-conversation-tracking<commit_after>from flask import Flask, request, session from twilio import twiml # The session object makes use of a secret key. SECRET_KEY = 'a secret key' app = Flask(__name__) app.config.from_object(__name__) # Try adding your own number to this list! callers = { "+14158675309": "Rey", "+14158675310": "Finn", "+14158675311": "Chewy", } @app.route("/", methods=['GET', 'POST']) def hello(): """Respond with the number of text messages sent between two parties.""" # Increment the counter counter = session.get('counter', 0) counter += 1 # Save the new counter value in the session session['counter'] = counter from_number = request.values.get('From') if from_number in callers: name = callers[from_number] else: name = "Friend" # Build our reply message = '{} has messaged {} {} times.' \ .format(name, request.values.get('To'), counter) # Put it in a TwiML response resp = twiml.Response() resp.message(message) return str(resp) if __name__ == "__main__": app.run(debug=True)
e6873d3d40d868e743a239c6a74a345c2999541e
dusty_coffin/elasticsearch_custom/edit_spreadsheet.py
dusty_coffin/elasticsearch_custom/edit_spreadsheet.py
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: print(d) res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
Edit spreadsheet is now fixed.
Edit spreadsheet is now fixed.
Python
mit
bhillmann/dusty_coffin,bhillmann/dusty_coffin,bhillmann/dusty_coffin,bhillmann/dusty_coffin
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created']) Edit spreadsheet is now fixed.
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: print(d) res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
<commit_before>import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created']) <commit_msg>Edit spreadsheet is now fixed.<commit_after>
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: print(d) res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created']) Edit spreadsheet is now fixed.import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: print(d) res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
<commit_before>import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created']) <commit_msg>Edit spreadsheet is now fixed.<commit_after>import urllib.request from io import BytesIO import pandas as pd import ujson from elasticsearch import Elasticsearch csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read() bio = BytesIO(csv) csv_pd = pd.DataFrame.from_csv(bio) json_objs = csv_pd.reset_index().to_json(orient='records') dict_array= ujson.loads(json_objs) # Edit to point to elasticsearch instance es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200') # ignore 404 and 400 es.indices.delete(index='datasets_index', ignore=[400, 404]) for d in dict_array: print(d) res = es.index(index="datasets_index", doc_type="dataset", body=d) print(res['created'])
8eb0b7fcd6ffb81d6b0fc69cb31c7625550583d7
targetrupypy.py
targetrupypy.py
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "rupypy-c" return entry_point, None def jitpolicy(driver): return JitPolicy()
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "./bin/topaz" return entry_point, None def jitpolicy(driver): return JitPolicy()
Move towards a normal bin directory.
Move towards a normal bin directory.
Python
bsd-3-clause
babelsberg/babelsberg-r,topazproject/topaz,babelsberg/babelsberg-r,kachick/topaz,kachick/topaz,babelsberg/babelsberg-r,babelsberg/babelsberg-r,babelsberg/babelsberg-r,kachick/topaz,topazproject/topaz,topazproject/topaz,topazproject/topaz
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "rupypy-c" return entry_point, None def jitpolicy(driver): return JitPolicy()Move towards a normal bin directory.
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "./bin/topaz" return entry_point, None def jitpolicy(driver): return JitPolicy()
<commit_before>from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "rupypy-c" return entry_point, None def jitpolicy(driver): return JitPolicy()<commit_msg>Move towards a normal bin directory.<commit_after>
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "./bin/topaz" return entry_point, None def jitpolicy(driver): return JitPolicy()
from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "rupypy-c" return entry_point, None def jitpolicy(driver): return JitPolicy()Move towards a normal bin directory.from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "./bin/topaz" return entry_point, None def jitpolicy(driver): return JitPolicy()
<commit_before>from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "rupypy-c" return entry_point, None def jitpolicy(driver): return JitPolicy()<commit_msg>Move towards a normal bin directory.<commit_after>from pypy.jit.codewriter.policy import JitPolicy from rupypy.main import entry_point def target(driver, args): driver.exe_name = "./bin/topaz" return entry_point, None def jitpolicy(driver): return JitPolicy()
39561a89ea497776d980d3eda97fc2f75493528f
internal_social_auth/views.py
internal_social_auth/views.py
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
Clean up the AuthComplete API a little
Clean up the AuthComplete API a little
Python
bsd-2-clause
incuna/incuna-internal-social-auth
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) Clean up the AuthComplete API a little
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
<commit_before>import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) <commit_msg>Clean up the AuthComplete API a little<commit_after>
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) Clean up the AuthComplete API a littleimport logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
<commit_before>import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, "Your Google Apps domain isn't authorized for this app") return HttpResponseRedirect('/') class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401) <commit_msg>Clean up the AuthComplete API a little<commit_after>import logging from django.contrib import messages from django.http import HttpResponseRedirect, HttpResponse from django.utils.encoding import force_text from django.views.generic.base import View from social_auth.exceptions import AuthFailed from social_auth.views import complete logger = logging.getLogger(__name__) class AuthComplete(View): def get(self, request, *args, **kwargs): backend = kwargs.pop('backend') try: return complete(request, backend, *args, **kwargs) except AuthFailed as e: logger.error(e) messages.error(request, self.get_error_message()) return HttpResponseRedirect(self.get_faiure_url()) def get_error_message(self): if self.error_message: return self.error_message return "Your Google Apps domain isn't authorized for this app" def get_failure_url(self): if self.failure_url: return force_text(self.failure_url) return '/' class LoginError(View): def get(self, request, *args, **kwargs): return HttpResponse(status=401)
97b000547898d6aa3006f07cf2ef9d8656a67865
conllu/__init__.py
conllu/__init__.py
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def _iter_sents(in_file): buf = [] for line in in_file: if line == "\n": yield "".join(buf)[:-1] buf = [] else: buf.append(line) if buf: yield "".join(buf) def parse_incr(in_file, fields=None): for sentence in _iter_sents(in_file): yield TokenList(*parse_token_and_metadata(sentence, fields=fields)) def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences def parse_tree_incr(in_file): for tokenlist in parse_incr(in_file): yield tokenlist.to_tree()
Add incremental parsing versions of parse and parse_tree
Add incremental parsing versions of parse and parse_tree
Python
mit
EmilStenstrom/conllu
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences Add incremental parsing versions of parse and parse_tree
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def _iter_sents(in_file): buf = [] for line in in_file: if line == "\n": yield "".join(buf)[:-1] buf = [] else: buf.append(line) if buf: yield "".join(buf) def parse_incr(in_file, fields=None): for sentence in _iter_sents(in_file): yield TokenList(*parse_token_and_metadata(sentence, fields=fields)) def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences def parse_tree_incr(in_file): for tokenlist in parse_incr(in_file): yield tokenlist.to_tree()
<commit_before>from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences <commit_msg>Add incremental parsing versions of parse and parse_tree<commit_after>
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def _iter_sents(in_file): buf = [] for line in in_file: if line == "\n": yield "".join(buf)[:-1] buf = [] else: buf.append(line) if buf: yield "".join(buf) def parse_incr(in_file, fields=None): for sentence in _iter_sents(in_file): yield TokenList(*parse_token_and_metadata(sentence, fields=fields)) def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences def parse_tree_incr(in_file): for tokenlist in parse_incr(in_file): yield tokenlist.to_tree()
from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences Add incremental parsing versions of parse and parse_treefrom __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def _iter_sents(in_file): buf = [] for line in in_file: if line == "\n": yield "".join(buf)[:-1] buf = [] else: buf.append(line) if buf: yield "".join(buf) def parse_incr(in_file, fields=None): for sentence in _iter_sents(in_file): yield TokenList(*parse_token_and_metadata(sentence, fields=fields)) def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences def parse_tree_incr(in_file): for tokenlist in parse_incr(in_file): yield tokenlist.to_tree()
<commit_before>from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences <commit_msg>Add incremental parsing versions of parse and parse_tree<commit_after>from __future__ import unicode_literals from conllu.models import TokenList from conllu.parser import parse_token_and_metadata def parse(data, fields=None): return [ TokenList(*parse_token_and_metadata(sentence, fields=fields)) for sentence in data.split("\n\n") if sentence ] def _iter_sents(in_file): buf = [] for line in in_file: if line == "\n": yield "".join(buf)[:-1] buf = [] else: buf.append(line) if buf: yield "".join(buf) def parse_incr(in_file, fields=None): for sentence in _iter_sents(in_file): yield TokenList(*parse_token_and_metadata(sentence, fields=fields)) def parse_tree(data): tokenlists = parse(data) sentences = [] for tokenlist in tokenlists: sentences.append(tokenlist.to_tree()) return sentences def parse_tree_incr(in_file): for tokenlist in parse_incr(in_file): yield tokenlist.to_tree()
9e3a6190b2dcfd7de03ef5c974b400a51219839e
pyof/v0x04/symmetric/hello.py
pyof/v0x04/symmetric/hello.py
"""Defines Hello message.""" # System imports # Third-party imports from pyof.v0x01.symmetric.hello import Hello __all__ = ('Hello',)
"""Defines Hello message.""" # System imports from enum import Enum from pyof.foundation.base import GenericMessage, GenericStruct from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16 from pyof.v0x04.common.header import Header, Type # Third-party imports __all__ = ('Hello', 'HelloElemHeader', 'HelloElemType', 'HelloElemVersionbitmap', 'ListOfHelloElements') # Enums class HelloElemType(Enum): """Hello element types.""" #: Bitmap of version supported. OFPHET_VERSIONBITMAP = 1 # Classes class HelloElemHeader(GenericStruct): """Common header for all Hello Elements.""" element_type = UBInt16(enum_ref=HelloElemType) length = UBInt16() def __init__(self, element_type=None, length=None): """The constructor just assigns parameters to object attributes. Args: element_type: One of OFPHET_*. length: Length in bytes of the element, including this header, excluding padding. """ super().__init__() self.element_type = element_type self.length = length class ListOfHelloElements(FixedTypeList): """List of Hello elements. Represented by instances of HelloElemHeader and used on Hello objects. """ def __init__(self, items=None): """The constructor just assigns parameters to object attributes. Args: items (HelloElemHeader): Instance or a list of instances. """ super().__init__(pyof_class=HelloElemHeader, items=items) class Hello(GenericMessage): """OpenFlow Hello Message OFPT_HELLO. This message includes zero or more hello elements having variable size. Unknown element types must be ignored/skipped, to allow for future extensions. """ header = Header(message_type=Type.OFPT_HELLO, length=8) #: Hello element list elements = ListOfHelloElements() def __init__(self, xid=None, elements=None): """The constructor takes the parameters below. Args: xid (int): xid to be used on the message header. elements: List of elements - 0 or more """ super().__init__(xid) self.elements = elements class HelloElemVersionbitmap(HelloElemHeader): """Version bitmap Hello Element.""" #: List of bitmaps - supported versions bitmaps = BinaryData() def __init__(self, bitmaps=b''): """The constructor just assigns parameters to object attributes. Args: bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then -Exactly (length + 7)/8*8 - (length) (between 0 and 7) bytes of all-zero bytes. """ super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP, length=None) self.bitmaps = bitmaps
Add Hello class and related classes for v0x04
Add Hello class and related classes for v0x04 Fix #302 Fix #303
Python
mit
cemsbr/python-openflow,kytos/python-openflow
"""Defines Hello message.""" # System imports # Third-party imports from pyof.v0x01.symmetric.hello import Hello __all__ = ('Hello',) Add Hello class and related classes for v0x04 Fix #302 Fix #303
"""Defines Hello message.""" # System imports from enum import Enum from pyof.foundation.base import GenericMessage, GenericStruct from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16 from pyof.v0x04.common.header import Header, Type # Third-party imports __all__ = ('Hello', 'HelloElemHeader', 'HelloElemType', 'HelloElemVersionbitmap', 'ListOfHelloElements') # Enums class HelloElemType(Enum): """Hello element types.""" #: Bitmap of version supported. OFPHET_VERSIONBITMAP = 1 # Classes class HelloElemHeader(GenericStruct): """Common header for all Hello Elements.""" element_type = UBInt16(enum_ref=HelloElemType) length = UBInt16() def __init__(self, element_type=None, length=None): """The constructor just assigns parameters to object attributes. Args: element_type: One of OFPHET_*. length: Length in bytes of the element, including this header, excluding padding. """ super().__init__() self.element_type = element_type self.length = length class ListOfHelloElements(FixedTypeList): """List of Hello elements. Represented by instances of HelloElemHeader and used on Hello objects. """ def __init__(self, items=None): """The constructor just assigns parameters to object attributes. Args: items (HelloElemHeader): Instance or a list of instances. """ super().__init__(pyof_class=HelloElemHeader, items=items) class Hello(GenericMessage): """OpenFlow Hello Message OFPT_HELLO. This message includes zero or more hello elements having variable size. Unknown element types must be ignored/skipped, to allow for future extensions. """ header = Header(message_type=Type.OFPT_HELLO, length=8) #: Hello element list elements = ListOfHelloElements() def __init__(self, xid=None, elements=None): """The constructor takes the parameters below. Args: xid (int): xid to be used on the message header. elements: List of elements - 0 or more """ super().__init__(xid) self.elements = elements class HelloElemVersionbitmap(HelloElemHeader): """Version bitmap Hello Element.""" #: List of bitmaps - supported versions bitmaps = BinaryData() def __init__(self, bitmaps=b''): """The constructor just assigns parameters to object attributes. Args: bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then -Exactly (length + 7)/8*8 - (length) (between 0 and 7) bytes of all-zero bytes. """ super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP, length=None) self.bitmaps = bitmaps
<commit_before>"""Defines Hello message.""" # System imports # Third-party imports from pyof.v0x01.symmetric.hello import Hello __all__ = ('Hello',) <commit_msg>Add Hello class and related classes for v0x04 Fix #302 Fix #303<commit_after>
"""Defines Hello message.""" # System imports from enum import Enum from pyof.foundation.base import GenericMessage, GenericStruct from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16 from pyof.v0x04.common.header import Header, Type # Third-party imports __all__ = ('Hello', 'HelloElemHeader', 'HelloElemType', 'HelloElemVersionbitmap', 'ListOfHelloElements') # Enums class HelloElemType(Enum): """Hello element types.""" #: Bitmap of version supported. OFPHET_VERSIONBITMAP = 1 # Classes class HelloElemHeader(GenericStruct): """Common header for all Hello Elements.""" element_type = UBInt16(enum_ref=HelloElemType) length = UBInt16() def __init__(self, element_type=None, length=None): """The constructor just assigns parameters to object attributes. Args: element_type: One of OFPHET_*. length: Length in bytes of the element, including this header, excluding padding. """ super().__init__() self.element_type = element_type self.length = length class ListOfHelloElements(FixedTypeList): """List of Hello elements. Represented by instances of HelloElemHeader and used on Hello objects. """ def __init__(self, items=None): """The constructor just assigns parameters to object attributes. Args: items (HelloElemHeader): Instance or a list of instances. """ super().__init__(pyof_class=HelloElemHeader, items=items) class Hello(GenericMessage): """OpenFlow Hello Message OFPT_HELLO. This message includes zero or more hello elements having variable size. Unknown element types must be ignored/skipped, to allow for future extensions. """ header = Header(message_type=Type.OFPT_HELLO, length=8) #: Hello element list elements = ListOfHelloElements() def __init__(self, xid=None, elements=None): """The constructor takes the parameters below. Args: xid (int): xid to be used on the message header. elements: List of elements - 0 or more """ super().__init__(xid) self.elements = elements class HelloElemVersionbitmap(HelloElemHeader): """Version bitmap Hello Element.""" #: List of bitmaps - supported versions bitmaps = BinaryData() def __init__(self, bitmaps=b''): """The constructor just assigns parameters to object attributes. Args: bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then -Exactly (length + 7)/8*8 - (length) (between 0 and 7) bytes of all-zero bytes. """ super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP, length=None) self.bitmaps = bitmaps
"""Defines Hello message.""" # System imports # Third-party imports from pyof.v0x01.symmetric.hello import Hello __all__ = ('Hello',) Add Hello class and related classes for v0x04 Fix #302 Fix #303"""Defines Hello message.""" # System imports from enum import Enum from pyof.foundation.base import GenericMessage, GenericStruct from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16 from pyof.v0x04.common.header import Header, Type # Third-party imports __all__ = ('Hello', 'HelloElemHeader', 'HelloElemType', 'HelloElemVersionbitmap', 'ListOfHelloElements') # Enums class HelloElemType(Enum): """Hello element types.""" #: Bitmap of version supported. OFPHET_VERSIONBITMAP = 1 # Classes class HelloElemHeader(GenericStruct): """Common header for all Hello Elements.""" element_type = UBInt16(enum_ref=HelloElemType) length = UBInt16() def __init__(self, element_type=None, length=None): """The constructor just assigns parameters to object attributes. Args: element_type: One of OFPHET_*. length: Length in bytes of the element, including this header, excluding padding. """ super().__init__() self.element_type = element_type self.length = length class ListOfHelloElements(FixedTypeList): """List of Hello elements. Represented by instances of HelloElemHeader and used on Hello objects. """ def __init__(self, items=None): """The constructor just assigns parameters to object attributes. Args: items (HelloElemHeader): Instance or a list of instances. """ super().__init__(pyof_class=HelloElemHeader, items=items) class Hello(GenericMessage): """OpenFlow Hello Message OFPT_HELLO. This message includes zero or more hello elements having variable size. Unknown element types must be ignored/skipped, to allow for future extensions. """ header = Header(message_type=Type.OFPT_HELLO, length=8) #: Hello element list elements = ListOfHelloElements() def __init__(self, xid=None, elements=None): """The constructor takes the parameters below. Args: xid (int): xid to be used on the message header. elements: List of elements - 0 or more """ super().__init__(xid) self.elements = elements class HelloElemVersionbitmap(HelloElemHeader): """Version bitmap Hello Element.""" #: List of bitmaps - supported versions bitmaps = BinaryData() def __init__(self, bitmaps=b''): """The constructor just assigns parameters to object attributes. Args: bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then -Exactly (length + 7)/8*8 - (length) (between 0 and 7) bytes of all-zero bytes. """ super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP, length=None) self.bitmaps = bitmaps
<commit_before>"""Defines Hello message.""" # System imports # Third-party imports from pyof.v0x01.symmetric.hello import Hello __all__ = ('Hello',) <commit_msg>Add Hello class and related classes for v0x04 Fix #302 Fix #303<commit_after>"""Defines Hello message.""" # System imports from enum import Enum from pyof.foundation.base import GenericMessage, GenericStruct from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16 from pyof.v0x04.common.header import Header, Type # Third-party imports __all__ = ('Hello', 'HelloElemHeader', 'HelloElemType', 'HelloElemVersionbitmap', 'ListOfHelloElements') # Enums class HelloElemType(Enum): """Hello element types.""" #: Bitmap of version supported. OFPHET_VERSIONBITMAP = 1 # Classes class HelloElemHeader(GenericStruct): """Common header for all Hello Elements.""" element_type = UBInt16(enum_ref=HelloElemType) length = UBInt16() def __init__(self, element_type=None, length=None): """The constructor just assigns parameters to object attributes. Args: element_type: One of OFPHET_*. length: Length in bytes of the element, including this header, excluding padding. """ super().__init__() self.element_type = element_type self.length = length class ListOfHelloElements(FixedTypeList): """List of Hello elements. Represented by instances of HelloElemHeader and used on Hello objects. """ def __init__(self, items=None): """The constructor just assigns parameters to object attributes. Args: items (HelloElemHeader): Instance or a list of instances. """ super().__init__(pyof_class=HelloElemHeader, items=items) class Hello(GenericMessage): """OpenFlow Hello Message OFPT_HELLO. This message includes zero or more hello elements having variable size. Unknown element types must be ignored/skipped, to allow for future extensions. """ header = Header(message_type=Type.OFPT_HELLO, length=8) #: Hello element list elements = ListOfHelloElements() def __init__(self, xid=None, elements=None): """The constructor takes the parameters below. Args: xid (int): xid to be used on the message header. elements: List of elements - 0 or more """ super().__init__(xid) self.elements = elements class HelloElemVersionbitmap(HelloElemHeader): """Version bitmap Hello Element.""" #: List of bitmaps - supported versions bitmaps = BinaryData() def __init__(self, bitmaps=b''): """The constructor just assigns parameters to object attributes. Args: bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then -Exactly (length + 7)/8*8 - (length) (between 0 and 7) bytes of all-zero bytes. """ super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP, length=None) self.bitmaps = bitmaps
66ae5304c81d74e8f30e9274c90d0f83766744d7
datamodel/nodes/printer.py
datamodel/nodes/printer.py
import sys from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): sys.stdout.write(self._context) return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
import sys import os from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): try: sys.stdout.write(str(self._context)) sys.stdout.write(os.linesep) except: pass return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
Make sure we only write chars to stdout
Make sure we only write chars to stdout
Python
apache-2.0
csparpa/robograph,csparpa/robograph
import sys from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): sys.stdout.write(self._context) return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._contextMake sure we only write chars to stdout
import sys import os from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): try: sys.stdout.write(str(self._context)) sys.stdout.write(os.linesep) except: pass return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
<commit_before>import sys from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): sys.stdout.write(self._context) return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context<commit_msg>Make sure we only write chars to stdout<commit_after>
import sys import os from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): try: sys.stdout.write(str(self._context)) sys.stdout.write(os.linesep) except: pass return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
import sys from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): sys.stdout.write(self._context) return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._contextMake sure we only write chars to stdoutimport sys import os from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): try: sys.stdout.write(str(self._context)) sys.stdout.write(os.linesep) except: pass return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
<commit_before>import sys from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): sys.stdout.write(self._context) return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context<commit_msg>Make sure we only write chars to stdout<commit_after>import sys import os from datamodel.base import node class ConsolePrinter(node.Node): """ This node prints on stdout its context and then returns it as output. """ def input(self, context): self._context = context def output(self): try: sys.stdout.write(str(self._context)) sys.stdout.write(os.linesep) except: pass return self._context def reset(self): del self._context class LogPrinter(node.Node): """ This node prints its context on a statically defined logger and then returns it as output """ def __init__(self, logger, loglevel, stringify=False, name=None): """ :param logger: any logging.Logger subtype :param loglevel: the log level :param stringify: try to cast to str the context before passing it to the logger :param name: name of this node """ node.Node.__init__(self, name=name) self._logger = logger self._loglevel = loglevel self._stringify = stringify def input(self, context): self._context = context def output(self): str_context = self._context if self._stringify: try: str_context = str(self._context) except: pass # oops... self._logger.log(self._loglevel, str_context) return self._context def reset(self): del self._context
f7153fd88f07f99181f790a93559efd585272f18
nuxeo-drive-client/tests/test_copy.py
nuxeo-drive-client/tests/test_copy.py
from tests.common_unit_test import UnitTestCase class TestCopy(UnitTestCase): def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
from tests.common_unit_test import RandomBug, UnitTestCase class TestCopy(UnitTestCase): @RandomBug('NXDRIVE-808', target='linux', repeat=5) def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
Add RandomBug for Linux on test_synchronize_remote_copy
NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy
Python
lgpl-2.1
ssdi-drive/nuxeo-drive,ssdi-drive/nuxeo-drive,ssdi-drive/nuxeo-drive
from tests.common_unit_test import UnitTestCase class TestCopy(UnitTestCase): def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.') NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy
from tests.common_unit_test import RandomBug, UnitTestCase class TestCopy(UnitTestCase): @RandomBug('NXDRIVE-808', target='linux', repeat=5) def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
<commit_before>from tests.common_unit_test import UnitTestCase class TestCopy(UnitTestCase): def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.') <commit_msg>NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy<commit_after>
from tests.common_unit_test import RandomBug, UnitTestCase class TestCopy(UnitTestCase): @RandomBug('NXDRIVE-808', target='linux', repeat=5) def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
from tests.common_unit_test import UnitTestCase class TestCopy(UnitTestCase): def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.') NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copyfrom tests.common_unit_test import RandomBug, UnitTestCase class TestCopy(UnitTestCase): @RandomBug('NXDRIVE-808', target='linux', repeat=5) def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
<commit_before>from tests.common_unit_test import UnitTestCase class TestCopy(UnitTestCase): def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.') <commit_msg>NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy<commit_after>from tests.common_unit_test import RandomBug, UnitTestCase class TestCopy(UnitTestCase): @RandomBug('NXDRIVE-808', target='linux', repeat=5) def test_synchronize_remote_copy(self): local = self.local_client_1 remote = self.remote_document_client_1 # Create a file and a folder in the remote root workspace remote.make_file('/', 'test.odt', 'Some content.') remote.make_folder('/', 'Test folder') # Launch ndrive and check synchronization self.engine_1.start() self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/')) self.assertTrue(local.exists('/Test folder')) self.assertTrue(local.exists('/test.odt')) # Copy the file to the folder remotely remote.copy('/test.odt', '/Test folder') # Launch ndrive and check synchronization self.wait_sync(wait_for_async=True) self.assertTrue(local.exists('/test.odt')) self.assertEqual(local.get_content('/test.odt'), 'Some content.') self.assertTrue(local.exists('/Test folder/test.odt')) self.assertEqual(local.get_content('/Test folder/test.odt'), 'Some content.')
0dc72761a3b4b17098633df27fdbb70058afe311
geotrek/signage/migrations/0013_auto_20200423_1255.py
geotrek/signage/migrations/0013_auto_20200423_1255.py
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), ]
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion def delete_force(apps, schema_editor): # We can't import Infrastructure models directly as it may be a newer # version than this migration expects. We use the historical version. Blade = apps.get_model('signage', 'Blade') for blade in Blade.objects.filter(deleted=True): blade.delete() class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), migrations.RunPython(delete_force), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), ]
Change order migration, user runpython instead
Change order migration, user runpython instead
Python
bsd-2-clause
makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), ] Change order migration, user runpython instead
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion def delete_force(apps, schema_editor): # We can't import Infrastructure models directly as it may be a newer # version than this migration expects. We use the historical version. Blade = apps.get_model('signage', 'Blade') for blade in Blade.objects.filter(deleted=True): blade.delete() class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), migrations.RunPython(delete_force), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), ]
<commit_before># Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), ] <commit_msg>Change order migration, user runpython instead<commit_after>
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion def delete_force(apps, schema_editor): # We can't import Infrastructure models directly as it may be a newer # version than this migration expects. We use the historical version. Blade = apps.get_model('signage', 'Blade') for blade in Blade.objects.filter(deleted=True): blade.delete() class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), migrations.RunPython(delete_force), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), ]
# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), ] Change order migration, user runpython instead# Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion def delete_force(apps, schema_editor): # We can't import Infrastructure models directly as it may be a newer # version than this migration expects. We use the historical version. Blade = apps.get_model('signage', 'Blade') for blade in Blade.objects.filter(deleted=True): blade.delete() class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), migrations.RunPython(delete_force), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), ]
<commit_before># Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), ] <commit_msg>Change order migration, user runpython instead<commit_after># Generated by Django 2.0.13 on 2020-04-23 12:55 from django.db import migrations, models import django.db.models.deletion def delete_force(apps, schema_editor): # We can't import Infrastructure models directly as it may be a newer # version than this migration expects. We use the historical version. Blade = apps.get_model('signage', 'Blade') for blade in Blade.objects.filter(deleted=True): blade.delete() class Migration(migrations.Migration): dependencies = [ ('signage', '0012_auto_20200406_1411'), ] operations = [ migrations.AlterField( model_name='line', name='blade', field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines', to='signage.Blade', verbose_name='Blade'), ), migrations.RunPython(delete_force), migrations.RemoveField( model_name='blade', name='deleted', ), migrations.RemoveField( model_name='blade', name='structure', ), migrations.RemoveField( model_name='line', name='structure', ), ]
b6d4baa9d30362a291567f078c1f93df7a63aeaa
waterbutler/providers/osfstorage/metadata.py
waterbutler/providers/osfstorage/metadata.py
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def full_path(self): return self.raw.get('fullPath') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
Return full path if it exists for OSF
Return full path if it exists for OSF
Python
apache-2.0
cosenal/waterbutler,Ghalko/waterbutler,CenterForOpenScience/waterbutler,icereval/waterbutler,Johnetordoff/waterbutler,TomBaxter/waterbutler,rdhyee/waterbutler,felliott/waterbutler,RCOSDP/waterbutler,rafaeldelucena/waterbutler,chrisseto/waterbutler,hmoco/waterbutler,kwierman/waterbutler
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], } Return full path if it exists for OSF
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def full_path(self): return self.raw.get('fullPath') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
<commit_before>from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], } <commit_msg>Return full path if it exists for OSF<commit_after>
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def full_path(self): return self.raw.get('fullPath') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], } Return full path if it exists for OSFfrom waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def full_path(self): return self.raw.get('fullPath') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
<commit_before>from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], } <commit_msg>Return full path if it exists for OSF<commit_after>from waterbutler.core import metadata class BaseOsfStorageMetadata: @property def provider(self): return 'osfstorage' class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] @property def modified(self): return self.raw.get('modified') @property def size(self): return self.raw.get('size') @property def full_path(self): return self.raw.get('fullPath') @property def content_type(self): return None @property def extra(self): return { key: self.raw[key] for key in ('version', 'downloads', 'fullPath') if key in self.raw } class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata): @property def name(self): return self.raw['name'] @property def path(self): return self.raw['path'] class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata): @property def modified(self): return self.raw['date'] @property def version_identifier(self): return 'version' @property def version(self): return str(self.raw['index']) @property def extra(self): return { 'user': self.raw['user'], 'downloads': self.raw['downloads'], }
f2fd526e08cc5576c651a7677c781c0c0bb7c94c
tests/test_jg.py
tests/test_jg.py
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') dot.render('t1.dot') assert dot.source == FIXTURE_GRAPH
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') assert dot.source == FIXTURE_GRAPH
Fix test generating graph file
Fix test generating graph file
Python
bsd-2-clause
abele/jinja-graph
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') dot.render('t1.dot') assert dot.source == FIXTURE_GRAPH Fix test generating graph file
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') assert dot.source == FIXTURE_GRAPH
<commit_before>from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') dot.render('t1.dot') assert dot.source == FIXTURE_GRAPH <commit_msg>Fix test generating graph file<commit_after>
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') assert dot.source == FIXTURE_GRAPH
from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') dot.render('t1.dot') assert dot.source == FIXTURE_GRAPH Fix test generating graph filefrom jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') assert dot.source == FIXTURE_GRAPH
<commit_before>from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') dot.render('t1.dot') assert dot.source == FIXTURE_GRAPH <commit_msg>Fix test generating graph file<commit_after>from jg.__main__ import main, generate_template_graph from mock import patch FIXTURE_GRAPH = ( 'digraph {\n' '\t"snippets/sub/analytics.html"\n' '\t"snippets/ga.html"\n' '\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n' '\t"header.html"\n' '\t"analytics.html"\n' '\t"custom_index.html"\n' '\t\t"custom_index.html" -> "index.html"\n' '\t\t"custom_index.html" -> "snippets/ga.html"\n' '\t"index.html"\n' '\t\t"index.html" -> "header.html"\n' '\t\t"index.html" -> "footer.html"\n' '\t"footer.html"\n}') def test_main_generates_graph_for_given_directory(): output_filename = 'graph.dot' with patch('jg.__main__.write') as write: exit_code = main(['./tests/fixtures', output_filename]) write.assert_called_with(FIXTURE_GRAPH, output_filename) assert exit_code == 0 def test_parses_all_templates_in_given_root_directory(): dot = generate_template_graph(root_path='./tests/fixtures') assert dot.source == FIXTURE_GRAPH
fe9a47f480b8db8de3b2b572f333e56497462ea2
Python/item15.py
Python/item15.py
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority2(num,pro): found=[False] def helper(x): nonlocal found if x in pro: found[0]=True return (0,x) return (1,x) num.sort(key=helper) return found def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority2(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
Add the sort_priority2 for python2.
Add the sort_priority2 for python2.
Python
mit
Vayne-Lover/Effective
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers) Add the sort_priority2 for python2.
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority2(num,pro): found=[False] def helper(x): nonlocal found if x in pro: found[0]=True return (0,x) return (1,x) num.sort(key=helper) return found def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority2(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
<commit_before># -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers) <commit_msg>Add the sort_priority2 for python2.<commit_after>
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority2(num,pro): found=[False] def helper(x): nonlocal found if x in pro: found[0]=True return (0,x) return (1,x) num.sort(key=helper) return found def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority2(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers) Add the sort_priority2 for python2.# -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority2(num,pro): found=[False] def helper(x): nonlocal found if x in pro: found[0]=True return (0,x) return (1,x) num.sort(key=helper) return found def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority2(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
<commit_before># -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers) <commit_msg>Add the sort_priority2 for python2.<commit_after># -*- coding: utf-8 -*- def sort_priority(num,pro): res=num[:] def helper(x): if x in pro: return (0,x) return (1,x) res.sort(key=helper) return res def sort_priority2(num,pro): found=[False] def helper(x): nonlocal found if x in pro: found[0]=True return (0,x) return (1,x) num.sort(key=helper) return found def sort_priority3(num,pro): found=False def helper(x): nonlocal found if x in pro: found=True return (0,x) return (1,x) num.sort(key=helper) return found if __name__=='__main__': numbers=[2,5,7,4,1,3,8,6] group=[2,4,8] print(sort_priority(numbers,group)) print(numbers) print(sort_priority2(numbers,group)) print(numbers) print(sort_priority3(numbers,group)) print(numbers)
8c7c2a65c9d9793e8ba4167ca629eda155f42c4e
mopidy/settings.py
mopidy/settings.py
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
Add logger name to log output
Add logger name to log output
Python
apache-2.0
quartz55/mopidy,bacontext/mopidy,mopidy/mopidy,mokieyue/mopidy,vrs01/mopidy,bencevans/mopidy,abarisain/mopidy,adamcik/mopidy,jmarsik/mopidy,tkem/mopidy,SuperStarPL/mopidy,jmarsik/mopidy,mokieyue/mopidy,ali/mopidy,dbrgn/mopidy,ZenithDK/mopidy,tkem/mopidy,priestd09/mopidy,adamcik/mopidy,woutervanwijk/mopidy,quartz55/mopidy,rawdlite/mopidy,liamw9534/mopidy,diandiankan/mopidy,adamcik/mopidy,hkariti/mopidy,rawdlite/mopidy,jodal/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,mokieyue/mopidy,vrs01/mopidy,jodal/mopidy,jmarsik/mopidy,tkem/mopidy,vrs01/mopidy,hkariti/mopidy,bencevans/mopidy,hkariti/mopidy,diandiankan/mopidy,priestd09/mopidy,abarisain/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,swak/mopidy,bacontext/mopidy,dbrgn/mopidy,pacificIT/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,bacontext/mopidy,quartz55/mopidy,bencevans/mopidy,jcass77/mopidy,ZenithDK/mopidy,jcass77/mopidy,kingosticks/mopidy,ZenithDK/mopidy,ali/mopidy,jmarsik/mopidy,rawdlite/mopidy,swak/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,dbrgn/mopidy,hkariti/mopidy,glogiotatidis/mopidy,ali/mopidy,liamw9534/mopidy,quartz55/mopidy,swak/mopidy,jodal/mopidy,bencevans/mopidy,swak/mopidy,vrs01/mopidy,mopidy/mopidy,dbrgn/mopidy,ZenithDK/mopidy,glogiotatidis/mopidy,mopidy/mopidy,mokieyue/mopidy,priestd09/mopidy,kingosticks/mopidy,rawdlite/mopidy,ali/mopidy,SuperStarPL/mopidy,kingosticks/mopidy,woutervanwijk/mopidy,jcass77/mopidy,tkem/mopidy,bacontext/mopidy,pacificIT/mopidy
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 Add logger name to log output
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
<commit_before>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 <commit_msg>Add logger name to log output<commit_after>
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 Add logger name to log outputCONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
<commit_before>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600 <commit_msg>Add logger name to log output<commit_after>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s' MPD_LINE_ENCODING = u'utf-8' MPD_LINE_TERMINATOR = u'\n' MPD_SERVER_HOSTNAME = u'localhost' MPD_SERVER_PORT = 6600
d4ea0bb24d79fa7900ffad4676a09cd4b830a4dc
parser_queuelog.py
parser_queuelog.py
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().split("\n") except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().splitlines() except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
Change split file lines by line
Change split file lines by line
Python
mit
roramirez/qpanel,roramirez/qpanel,roramirez/qpanel,skazancev/qpanel,skazancev/qpanel,skazancev/qpanel,skazancev/qpanel,roramirez/qpanel
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().split("\n") except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse() Change split file lines by line
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().splitlines() except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
<commit_before># -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().split("\n") except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse() <commit_msg>Change split file lines by line<commit_after>
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().splitlines() except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().split("\n") except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse() Change split file lines by line# -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().splitlines() except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
<commit_before># -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().split("\n") except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse() <commit_msg>Change split file lines by line<commit_after># -*- coding: utf-8 -*- # # Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com> # # Parse queue_log Asterisk file and add records into database. # from libs.qpanel import model import click import sys @click.command() @click.option('--file', default='/var/log/asterisk/queue_log', help='Queue Log file.') @click.option('--verbose', default=False) def parse(file, verbose): inserted, not_inserted = 0, 0 try: with open(file) as fb: print("Reading file %s ..." % file) content = fb.read().splitlines() except IOError: print('File file %s not exits or not can read.' % file) sys.exit(1) for idx, line in enumerate(content): record = line.split('|') if len(record) < 4: continue if not exist_record(record) and insert_record(record): inserted += 1 if verbose: print ("Insert record ", record) else: if verbose: print ("Not insert record ", record) not_inserted += 1 print ("Insert record: %i\nNo inserted record: %i" % (inserted, not_inserted)) def exist_record(record): return model.queuelog_exists_record(record) def insert_record(record): return model.queuelog_insert(record) if __name__ == '__main__': parse()
0599e76db6c1eef495a608d7386601bbee3cfbc5
test/authinfo.py
test/authinfo.py
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Empty Result with HTTP GET ''' def testEmptyInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
Add test for empty responses
Add test for empty responses
Python
mit
fbraem/mqweb,fbraem/mqweb,fbraem/mqweb
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite) Add test for empty responses
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Empty Result with HTTP GET ''' def testEmptyInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
<commit_before>import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite) <commit_msg>Add test for empty responses<commit_after>
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Empty Result with HTTP GET ''' def testEmptyInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite) Add test for empty responsesimport unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Empty Result with HTTP GET ''' def testEmptyInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
<commit_before>import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite) <commit_msg>Add test for empty responses<commit_after>import unittest from testbase import MQWebTest ''' Test for AuthenticationInformationController ''' class TestAuthInfoActions(MQWebTest): ''' Test Inquire with HTTP GET ''' def testInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr) ''' Test Empty Result with HTTP GET ''' def testEmptyInquire(self): json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True) ''' Test Inquire with HTTP POST ''' def testInquireJSON(self): print filter = { "AuthInfoName" : "*" } json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter) suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions) unittest.TextTestRunner(verbosity=2).run(suite)
d4a0a85673b5d61b82c65e77efcd6518da719952
pmxbot/__init__.py
pmxbot/__init__.py
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
Remove places default config. It doesn't appear to be used anywhere.
Remove places default config. It doesn't appear to be used anywhere.
Python
mit
yougov/pmxbot,yougov/pmxbot,yougov/pmxbot
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object" Remove places default config. It doesn't appear to be used anywhere.
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
<commit_before># vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object" <commit_msg>Remove places default config. It doesn't appear to be used anywhere.<commit_after>
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object" Remove places default config. It doesn't appear to be used anywhere.# vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
<commit_before># vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], places=['London', 'Tokyo', 'New York'], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object" <commit_msg>Remove places default config. It doesn't appear to be used anywhere.<commit_after># vim:ts=4:sw=4:noexpandtab import socket import logging as _logging from .dictlib import ConfigDict config = ConfigDict( bot_nickname='pmxbot', database='sqlite:pmxbot.sqlite', server_host='localhost', server_port=6667, use_ssl=False, password=None, nickserv_password=None, silent_bot=False, log_channels=[], other_channels=[], librarypaste='http://paste.jaraco.com', ) config['logs URL'] = 'http://' + socket.getfqdn() config['log level'] = _logging.INFO "The config object"
525cbab46570342098613ae591749b4cf5026453
tests/terrain.py
tests/terrain.py
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir)
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir) # Create output directory if not exists if not os.path.exists("test_output"): os.makedirs("test_output")
Create the tests output directory automatically.
Create the tests output directory automatically.
Python
mit
gnott/elife-poa-xml-generation,gnott/elife-poa-xml-generation
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir)Create the tests output directory automatically.
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir) # Create output directory if not exists if not os.path.exists("test_output"): os.makedirs("test_output")
<commit_before>from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir)<commit_msg>Create the tests output directory automatically.<commit_after>
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir) # Create output directory if not exists if not os.path.exists("test_output"): os.makedirs("test_output")
from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir)Create the tests output directory automatically.from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir) # Create output directory if not exists if not os.path.exists("test_output"): os.makedirs("test_output")
<commit_before>from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir)<commit_msg>Create the tests output directory automatically.<commit_after>from lettuce import world import os """ Set world.basedir relative to this terrain.py file, when running lettuce from this directory, and add the directory it to the import path """ world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) os.sys.path.insert(0,world.basedir) world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__))) os.sys.path.insert(0,world.basedir) # Create output directory if not exists if not os.path.exists("test_output"): os.makedirs("test_output")
8d09a8557433d95015010465b62f31ffe7b6fe2c
usb/shortener.py
usb/shortener.py
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = None def get_short_id(self, number): if self._hasher is None: self._hasher = Hashids(self.secret, self.min_length) return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = Hashids(self.secret, self.min_length) def get_short_id(self, number): return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
Move hasher creation to constructor
Move hasher creation to constructor
Python
mit
dizpers/usb
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = None def get_short_id(self, number): if self._hasher is None: self._hasher = Hashids(self.secret, self.min_length) return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}' Move hasher creation to constructor
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = Hashids(self.secret, self.min_length) def get_short_id(self, number): return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
<commit_before>from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = None def get_short_id(self, number): if self._hasher is None: self._hasher = Hashids(self.secret, self.min_length) return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}' <commit_msg>Move hasher creation to constructor<commit_after>
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = Hashids(self.secret, self.min_length) def get_short_id(self, number): return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = None def get_short_id(self, number): if self._hasher is None: self._hasher = Hashids(self.secret, self.min_length) return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}' Move hasher creation to constructorfrom hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = Hashids(self.secret, self.min_length) def get_short_id(self, number): return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
<commit_before>from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = None def get_short_id(self, number): if self._hasher is None: self._hasher = Hashids(self.secret, self.min_length) return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}' <commit_msg>Move hasher creation to constructor<commit_after>from hashids import Hashids class Shortener(object): def __init__(self, secret, min_length, short_url_domain): self.secret = secret self.min_length = min_length self.short_url_domain = short_url_domain self._hasher = Hashids(self.secret, self.min_length) def get_short_id(self, number): return self._hasher.encode(number) def get_short_url(self, short_id): return f'http://{self.short_url_domain}/{short_id}'
b5ae6290382ef69f9d76c0494aee90f85bdf2c16
plugins/Views/SimpleView/__init__.py
plugins/Views/SimpleView/__init__.py
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "type": "view", "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
Fix plug-in type and description key
Fix plug-in type and description key 't Was a typo. Contributes to issue CURA-1190.
Python
agpl-3.0
onitake/Uranium,onitake/Uranium
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() } Fix plug-in type and description key 't Was a typo. Contributes to issue CURA-1190.
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "type": "view", "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
<commit_before># Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() } <commit_msg>Fix plug-in type and description key 't Was a typo. Contributes to issue CURA-1190.<commit_after>
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "type": "view", "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() } Fix plug-in type and description key 't Was a typo. Contributes to issue CURA-1190.# Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "type": "view", "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
<commit_before># Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() } <commit_msg>Fix plug-in type and description key 't Was a typo. Contributes to issue CURA-1190.<commit_after># Copyright (c) 2015 Ultimaker B.V. # Uranium is released under the terms of the AGPLv3 or higher. from . import SimpleView from UM.i18n import i18nCatalog i18n_catalog = i18nCatalog("uranium") def getMetaData(): return { "type": "view", "plugin": { "name": i18n_catalog.i18nc("@label", "Simple View"), "author": "Ultimaker", "version": "1.0", "description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."), "api": 2 }, "view": { "name": i18n_catalog.i18nc("@item:inmenu", "Simple"), "visible": False } } def register(app): return { "view": SimpleView.SimpleView() }
35c7d7816c3c441286519658a3426a5f03aca284
plugins/check_pinned/check_pinned.py
plugins/check_pinned/check_pinned.py
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE")
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") # Can we send a message to the Slackbot with the UserID # (which we have) and have the slackbot post to the user?
Add note about potential work around
Add note about potential work around The bot itself cannot send a message only on presence_change actions We may be able to contact the slack bot to send a message in our stead instead.
Python
mit
pyamanak/oithdbot
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") Add note about potential work around The bot itself cannot send a message only on presence_change actions We may be able to contact the slack bot to send a message in our stead instead.
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") # Can we send a message to the Slackbot with the UserID # (which we have) and have the slackbot post to the user?
<commit_before>from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") <commit_msg>Add note about potential work around The bot itself cannot send a message only on presence_change actions We may be able to contact the slack bot to send a message in our stead instead.<commit_after>
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") # Can we send a message to the Slackbot with the UserID # (which we have) and have the slackbot post to the user?
from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") Add note about potential work around The bot itself cannot send a message only on presence_change actions We may be able to contact the slack bot to send a message in our stead instead.from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") # Can we send a message to the Slackbot with the UserID # (which we have) and have the slackbot post to the user?
<commit_before>from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") <commit_msg>Add note about potential work around The bot itself cannot send a message only on presence_change actions We may be able to contact the slack bot to send a message in our stead instead.<commit_after>from __future__ import unicode_literals # don't convert to ascii in py2.7 when creating string to return crontable = [] outputs = [] # Catch all the events def catch_all(data): print(data) # Only handles when a user becomes active def process_presence_change(data): print("PRESENCE CHANGE") # While we can respond to presence change events, # we cannot actually send a message to a channel as # the data structure does not contain a channel ID if (data["presence"].startswith("active")): print("IS ACTIVE") # Can we send a message to the Slackbot with the UserID # (which we have) and have the slackbot post to the user?
e4ee7034291fbeda48efa0d1c617be8a20eb49bd
algorithms/python/496_next_greater_element.py
algorithms/python/496_next_greater_element.py
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
Add another solution for 496 next greater element
Add another solution for 496 next greater element
Python
mit
ruichao-factual/leetcode
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results Add another solution for 496 next greater element
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
<commit_before>class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results <commit_msg>Add another solution for 496 next greater element<commit_after>
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results Add another solution for 496 next greater elementclass Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
<commit_before>class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results <commit_msg>Add another solution for 496 next greater element<commit_after>class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ results = [] for findNum in findNums: index = nums.index(findNum) result = index + 1 for candidate in nums[index + 1:]: if candidate > findNum: results.append(candidate) break else: result += 1 if result >= len(nums): results.append(-1) return results # Solution 2 class Solution(object): def nextGreaterElement(self, findNums, nums): """ :type findNums: List[int] :type nums: List[int] :rtype: List[int] """ result_hash = {} stack = [] for num in nums: while stack and num > stack[-1]: result_hash[stack.pop()] = num stack.append(num) return [result_hash.get(x, -1) for x in findNums]
c480c2058d7f5d2e9bfe2e1a6f0a318eb0d49a4b
tangled/website/resources.py
tangled/website/resources.py
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template_name='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
Rename template_name @config arg to template
Rename template_name @config arg to template
Python
mit
TangledWeb/tangled.website
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template_name='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data Rename template_name @config arg to template
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
<commit_before>from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template_name='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data <commit_msg>Rename template_name @config arg to template<commit_after>
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template_name='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data Rename template_name @config arg to templatefrom tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
<commit_before>from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template_name='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data <commit_msg>Rename template_name @config arg to template<commit_after>from tangled.web import Resource, config from tangled.site.resources.entry import Entry class Docs(Entry): @config('text/html', template='tangled.website:templates/docs.mako') def GET(self): static_dirs = self.app.get_all('static_directory', as_dict=True) links = [] for prefix, dir_app in static_dirs.items(): if prefix[0] == 'docs': links.append({ 'href': '/'.join(prefix), 'text': prefix[1], }) self.urlvars['id'] = 'docs' data = super().GET() data['links'] = sorted(links, key=lambda i: i['text']) return data
428371a67a7c742e239d96914a96558171f9f59e
test/requests/test-website.py
test/requests/test-website.py
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules import requests as req import sys print "Mechanical Rob firing up..." if len(sys.argv)<1: raise "Problem with arguments" url = sys.argv[1] print url r = req.get(url) print r
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules from __future__ import print_function from link_checker import check_links import argparse print("Mechanical Rob firing up...") def run_all(args_obj, parser): print("") print("Running all tests.") check_links(args_obj, parser) # TODO: Add other functions as they are created. def print_help(args_obj, parser): print(parser.format_help()) def dummy(args_obj, parser): print("Not implemented yet.") desc = """ This is Mechanical-Rob - an automated web server tester for Genenetwork.org """ parser = argparse.ArgumentParser(description=desc) parser.add_argument("-d", "--database", metavar="DB", type=str , default="db_webqtl_s" , help="Use database (default db_webqtl_s)") parser.add_argument("host", metavar="HOST", type=str , default="http://localhost:5003" , help="The url to the web server") parser.add_argument("-a", "--all", dest="accumulate", action="store_const" , const=run_all, default=print_help , help="Runs all tests.") parser.add_argument("-l", "--link-checker", dest="accumulate" , action='store_const', const=check_links, default=print_help , help="Checks for dead links.") # parser.add_argument("-n", "--navigation", dest="accumulate" # , action="store_const", const=check_navigation, default=print_help # , help="Checks for navigation.") # parser.add_argument("-m", "--mapping", dest="accumulate" # , action="store_const", const=check_mapping, default=print_help # , help="Checks for mapping.") # parser.add_argument("-s", "--skip-broken", dest="accumulate" # , action="store_const", const=dummy, default=print_help # , help="Skip tests that are known to be broken.") args = parser.parse_args() # print("The arguments object: ", args) args.accumulate(args, parser)
Use argparse to handle arguments
Use argparse to handle arguments * Use argparse to handle commandline arguments. * Create initial layout of how the code might end up - lots of the code is currently commented out.
Python
agpl-3.0
zsloan/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,genenetwork/genenetwork2,pjotrp/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,zsloan/genenetwork2
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules import requests as req import sys print "Mechanical Rob firing up..." if len(sys.argv)<1: raise "Problem with arguments" url = sys.argv[1] print url r = req.get(url) print r Use argparse to handle arguments * Use argparse to handle commandline arguments. * Create initial layout of how the code might end up - lots of the code is currently commented out.
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules from __future__ import print_function from link_checker import check_links import argparse print("Mechanical Rob firing up...") def run_all(args_obj, parser): print("") print("Running all tests.") check_links(args_obj, parser) # TODO: Add other functions as they are created. def print_help(args_obj, parser): print(parser.format_help()) def dummy(args_obj, parser): print("Not implemented yet.") desc = """ This is Mechanical-Rob - an automated web server tester for Genenetwork.org """ parser = argparse.ArgumentParser(description=desc) parser.add_argument("-d", "--database", metavar="DB", type=str , default="db_webqtl_s" , help="Use database (default db_webqtl_s)") parser.add_argument("host", metavar="HOST", type=str , default="http://localhost:5003" , help="The url to the web server") parser.add_argument("-a", "--all", dest="accumulate", action="store_const" , const=run_all, default=print_help , help="Runs all tests.") parser.add_argument("-l", "--link-checker", dest="accumulate" , action='store_const', const=check_links, default=print_help , help="Checks for dead links.") # parser.add_argument("-n", "--navigation", dest="accumulate" # , action="store_const", const=check_navigation, default=print_help # , help="Checks for navigation.") # parser.add_argument("-m", "--mapping", dest="accumulate" # , action="store_const", const=check_mapping, default=print_help # , help="Checks for mapping.") # parser.add_argument("-s", "--skip-broken", dest="accumulate" # , action="store_const", const=dummy, default=print_help # , help="Skip tests that are known to be broken.") args = parser.parse_args() # print("The arguments object: ", args) args.accumulate(args, parser)
<commit_before># Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules import requests as req import sys print "Mechanical Rob firing up..." if len(sys.argv)<1: raise "Problem with arguments" url = sys.argv[1] print url r = req.get(url) print r <commit_msg>Use argparse to handle arguments * Use argparse to handle commandline arguments. * Create initial layout of how the code might end up - lots of the code is currently commented out.<commit_after>
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules from __future__ import print_function from link_checker import check_links import argparse print("Mechanical Rob firing up...") def run_all(args_obj, parser): print("") print("Running all tests.") check_links(args_obj, parser) # TODO: Add other functions as they are created. def print_help(args_obj, parser): print(parser.format_help()) def dummy(args_obj, parser): print("Not implemented yet.") desc = """ This is Mechanical-Rob - an automated web server tester for Genenetwork.org """ parser = argparse.ArgumentParser(description=desc) parser.add_argument("-d", "--database", metavar="DB", type=str , default="db_webqtl_s" , help="Use database (default db_webqtl_s)") parser.add_argument("host", metavar="HOST", type=str , default="http://localhost:5003" , help="The url to the web server") parser.add_argument("-a", "--all", dest="accumulate", action="store_const" , const=run_all, default=print_help , help="Runs all tests.") parser.add_argument("-l", "--link-checker", dest="accumulate" , action='store_const', const=check_links, default=print_help , help="Checks for dead links.") # parser.add_argument("-n", "--navigation", dest="accumulate" # , action="store_const", const=check_navigation, default=print_help # , help="Checks for navigation.") # parser.add_argument("-m", "--mapping", dest="accumulate" # , action="store_const", const=check_mapping, default=print_help # , help="Checks for mapping.") # parser.add_argument("-s", "--skip-broken", dest="accumulate" # , action="store_const", const=dummy, default=print_help # , help="Skip tests that are known to be broken.") args = parser.parse_args() # print("The arguments object: ", args) args.accumulate(args, parser)
# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules import requests as req import sys print "Mechanical Rob firing up..." if len(sys.argv)<1: raise "Problem with arguments" url = sys.argv[1] print url r = req.get(url) print r Use argparse to handle arguments * Use argparse to handle commandline arguments. * Create initial layout of how the code might end up - lots of the code is currently commented out.# Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules from __future__ import print_function from link_checker import check_links import argparse print("Mechanical Rob firing up...") def run_all(args_obj, parser): print("") print("Running all tests.") check_links(args_obj, parser) # TODO: Add other functions as they are created. def print_help(args_obj, parser): print(parser.format_help()) def dummy(args_obj, parser): print("Not implemented yet.") desc = """ This is Mechanical-Rob - an automated web server tester for Genenetwork.org """ parser = argparse.ArgumentParser(description=desc) parser.add_argument("-d", "--database", metavar="DB", type=str , default="db_webqtl_s" , help="Use database (default db_webqtl_s)") parser.add_argument("host", metavar="HOST", type=str , default="http://localhost:5003" , help="The url to the web server") parser.add_argument("-a", "--all", dest="accumulate", action="store_const" , const=run_all, default=print_help , help="Runs all tests.") parser.add_argument("-l", "--link-checker", dest="accumulate" , action='store_const', const=check_links, default=print_help , help="Checks for dead links.") # parser.add_argument("-n", "--navigation", dest="accumulate" # , action="store_const", const=check_navigation, default=print_help # , help="Checks for navigation.") # parser.add_argument("-m", "--mapping", dest="accumulate" # , action="store_const", const=check_mapping, default=print_help # , help="Checks for mapping.") # parser.add_argument("-s", "--skip-broken", dest="accumulate" # , action="store_const", const=dummy, default=print_help # , help="Skip tests that are known to be broken.") args = parser.parse_args() # print("The arguments object: ", args) args.accumulate(args, parser)
<commit_before># Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules import requests as req import sys print "Mechanical Rob firing up..." if len(sys.argv)<1: raise "Problem with arguments" url = sys.argv[1] print url r = req.get(url) print r <commit_msg>Use argparse to handle arguments * Use argparse to handle commandline arguments. * Create initial layout of how the code might end up - lots of the code is currently commented out.<commit_after># Run with something like # # env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003 # # Mostly to pick up the Guix GN2_PROFILE and python modules from __future__ import print_function from link_checker import check_links import argparse print("Mechanical Rob firing up...") def run_all(args_obj, parser): print("") print("Running all tests.") check_links(args_obj, parser) # TODO: Add other functions as they are created. def print_help(args_obj, parser): print(parser.format_help()) def dummy(args_obj, parser): print("Not implemented yet.") desc = """ This is Mechanical-Rob - an automated web server tester for Genenetwork.org """ parser = argparse.ArgumentParser(description=desc) parser.add_argument("-d", "--database", metavar="DB", type=str , default="db_webqtl_s" , help="Use database (default db_webqtl_s)") parser.add_argument("host", metavar="HOST", type=str , default="http://localhost:5003" , help="The url to the web server") parser.add_argument("-a", "--all", dest="accumulate", action="store_const" , const=run_all, default=print_help , help="Runs all tests.") parser.add_argument("-l", "--link-checker", dest="accumulate" , action='store_const', const=check_links, default=print_help , help="Checks for dead links.") # parser.add_argument("-n", "--navigation", dest="accumulate" # , action="store_const", const=check_navigation, default=print_help # , help="Checks for navigation.") # parser.add_argument("-m", "--mapping", dest="accumulate" # , action="store_const", const=check_mapping, default=print_help # , help="Checks for mapping.") # parser.add_argument("-s", "--skip-broken", dest="accumulate" # , action="store_const", const=dummy, default=print_help # , help="Skip tests that are known to be broken.") args = parser.parse_args() # print("The arguments object: ", args) args.accumulate(args, parser)
e5daa53aab94360c2e06a6cb608c4992b25becc6
test/helpers.py
test/helpers.py
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(1) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(1) f.data = data f.flags = flags return f
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None, stream_id=1): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(stream_id) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None, stream_id=1): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(stream_id) f.data = data f.flags = flags return f
Allow stream ids != 1 in frame factory.
Allow stream ids != 1 in frame factory.
Python
mit
Kriechi/hyper-h2,mhils/hyper-h2,bhavishyagopesh/hyper-h2,Kriechi/hyper-h2,vladmunteanu/hyper-h2,python-hyper/hyper-h2,python-hyper/hyper-h2,vladmunteanu/hyper-h2
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(1) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(1) f.data = data f.flags = flags return f Allow stream ids != 1 in frame factory.
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None, stream_id=1): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(stream_id) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None, stream_id=1): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(stream_id) f.data = data f.flags = flags return f
<commit_before># -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(1) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(1) f.data = data f.flags = flags return f <commit_msg>Allow stream ids != 1 in frame factory.<commit_after>
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None, stream_id=1): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(stream_id) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None, stream_id=1): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(stream_id) f.data = data f.flags = flags return f
# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(1) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(1) f.data = data f.flags = flags return f Allow stream ids != 1 in frame factory.# -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None, stream_id=1): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(stream_id) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None, stream_id=1): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(stream_id) f.data = data f.flags = flags return f
<commit_before># -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(1) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(1) f.data = data f.flags = flags return f <commit_msg>Allow stream ids != 1 in frame factory.<commit_after># -*- coding: utf-8 -*- """ helpers ~~~~~~~ This module contains helpers for the h2 tests. """ from hyperframe.frame import HeadersFrame, DataFrame from hpack.hpack import Encoder class FrameFactory(object): """ A class containing lots of helper methods and state to build frames. This allows test cases to easily build correct HTTP/2 frames to feed to hyper-h2. """ def __init__(self): self.encoder = Encoder() def preamble(self): return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n' def build_headers_frame(self, headers, flags=None, stream_id=1): """ Builds a single valid headers frame out of the contained headers. """ f = HeadersFrame(stream_id) f.data = self.encoder.encode(headers) f.flags.add('END_HEADERS') if flags: f.flags.update(flags) return f def build_data_frame(self, data, flags=None, stream_id=1): """ Builds a single data frame out of a chunk of data. """ flags = set(flags) if flags is not None else set() f = DataFrame(stream_id) f.data = data f.flags = flags return f
1509336a27d80eae68e56cfa776bd8342221297f
tests/scoring_engine/engine/test_basic_check.py
tests/scoring_engine/engine/test_basic_check.py
from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus'
import pytest from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' def test_check_no_properties(self): check = BasicCheck(self.environment) check.required_properties = ['testparam'] with pytest.raises(LookupError): check.set_properties()
Add test for incorrect check properties
Add test for incorrect check properties
Python
mit
pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine
from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' Add test for incorrect check properties
import pytest from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' def test_check_no_properties(self): check = BasicCheck(self.environment) check.required_properties = ['testparam'] with pytest.raises(LookupError): check.set_properties()
<commit_before>from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' <commit_msg>Add test for incorrect check properties<commit_after>
import pytest from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' def test_check_no_properties(self): check = BasicCheck(self.environment) check.required_properties = ['testparam'] with pytest.raises(LookupError): check.set_properties()
from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' Add test for incorrect check propertiesimport pytest from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' def test_check_no_properties(self): check = BasicCheck(self.environment) check.required_properties = ['testparam'] with pytest.raises(LookupError): check.set_properties()
<commit_before>from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' <commit_msg>Add test for incorrect check properties<commit_after>import pytest from scoring_engine.engine.basic_check import BasicCheck from scoring_engine.models.service import Service from scoring_engine.models.environment import Environment from scoring_engine.models.account import Account from tests.scoring_engine.unit_test import UnitTest class TestBasicCheck(UnitTest): def setup(self): super(TestBasicCheck, self).setup() self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1') self.environment = Environment(matching_regex='*', service=self.service) def test_init(self): check = BasicCheck(self.environment) assert check.environment == self.environment assert check.required_properties == [] def test_get_host(self): self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.host == '127.0.0.1' def test_get_random_account(self): self.session.add(Account(username='pwnbus', password='pass', service=self.service)) self.session.add(self.service) self.session.add(self.environment) self.session.commit() check = BasicCheck(self.environment) assert check.get_random_account().username == 'pwnbus' def test_check_no_properties(self): check = BasicCheck(self.environment) check.required_properties = ['testparam'] with pytest.raises(LookupError): check.set_properties()
8b5ccf93fbac8929ecfc185d7407a79b1e890bde
project_template/project_settings.py
project_template/project_settings.py
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings.
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
Use GLAMkit settings in default project template.
Use GLAMkit settings in default project template.
Python
mit
ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings. Use GLAMkit settings in default project template.
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
<commit_before># Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings. <commit_msg>Use GLAMkit settings in default project template.<commit_after>
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings. Use GLAMkit settings in default project template.# Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
<commit_before># Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.icekit import * # icekit, glamkit # Override the default ICEkit settings to form project settings. <commit_msg>Use GLAMkit settings in default project template.<commit_after># Do not commit secrets to VCS. # Local environment variables will be loaded from `.env.local`. # Additional environment variables will be loaded from `.env.$DOTENV`. # Local settings will be imported from `icekit_settings_local.py` from icekit.project.settings.glamkit import * # glamkit, icekit # Override the default ICEkit settings to form project settings.
ea20424f2645f1157c0a7ea76b32a4f834c1e27b
pythonect/internal/__init__.py
pythonect/internal/__init__.py
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
Delete blank line at end of file
Delete blank line at end of file
Python
bsd-3-clause
ikotler/pythonect,fr34k8/pythonect,fr34k8/pythonect,TOSPIO/pythonect,ikotler/pythonect
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. Delete blank line at end of file
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
<commit_before># Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. <commit_msg>Delete blank line at end of file<commit_after>
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. Delete blank line at end of file# Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
<commit_before># Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. <commit_msg>Delete blank line at end of file<commit_after># Copyright (c) 2012-2013, Itzik Kotler # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of the author nor the names of its contributors may # be used to endorse or promote products derived from this software without # specific prior written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY # DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
6cfc784ce3136cbec8c88948f4d6b45f9070b91b
pyqode/__init__.py
pyqode/__init__.py
# -*- coding: utf-8 -*- """ pyQode is a code editor framework for python qt applications. """ import pkg_resources pkg_resources.declare_namespace(__name__)
# -*- coding: utf-8 -*- """ pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide) pyQode is a **namespace package**. """ import pkg_resources pkg_resources.declare_namespace(__name__)
Fix pyqode main docstring (uniformised with other pyqode namespace packages)
Fix pyqode main docstring (uniformised with other pyqode namespace packages)
Python
mit
mmolero/pyqode.python,pyQode/pyqode.python,zwadar/pyqode.python,pyQode/pyqode.python
# -*- coding: utf-8 -*- """ pyQode is a code editor framework for python qt applications. """ import pkg_resources pkg_resources.declare_namespace(__name__) Fix pyqode main docstring (uniformised with other pyqode namespace packages)
# -*- coding: utf-8 -*- """ pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide) pyQode is a **namespace package**. """ import pkg_resources pkg_resources.declare_namespace(__name__)
<commit_before># -*- coding: utf-8 -*- """ pyQode is a code editor framework for python qt applications. """ import pkg_resources pkg_resources.declare_namespace(__name__) <commit_msg>Fix pyqode main docstring (uniformised with other pyqode namespace packages)<commit_after>
# -*- coding: utf-8 -*- """ pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide) pyQode is a **namespace package**. """ import pkg_resources pkg_resources.declare_namespace(__name__)
# -*- coding: utf-8 -*- """ pyQode is a code editor framework for python qt applications. """ import pkg_resources pkg_resources.declare_namespace(__name__) Fix pyqode main docstring (uniformised with other pyqode namespace packages)# -*- coding: utf-8 -*- """ pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide) pyQode is a **namespace package**. """ import pkg_resources pkg_resources.declare_namespace(__name__)
<commit_before># -*- coding: utf-8 -*- """ pyQode is a code editor framework for python qt applications. """ import pkg_resources pkg_resources.declare_namespace(__name__) <commit_msg>Fix pyqode main docstring (uniformised with other pyqode namespace packages)<commit_after># -*- coding: utf-8 -*- """ pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide) pyQode is a **namespace package**. """ import pkg_resources pkg_resources.declare_namespace(__name__)
226e8c322670a310fcfb9eb95d9d59838bbac3d3
refcollections/admin_custom.py
refcollections/admin_custom.py
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User shells_admin.register(User, UserAdmin)
Add User back into admin
Add User back into admin
Python
bsd-3-clause
uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) Add User back into admin
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User shells_admin.register(User, UserAdmin)
<commit_before>from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) <commit_msg>Add User back into admin<commit_after>
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User shells_admin.register(User, UserAdmin)
from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) Add User back into adminfrom django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User shells_admin.register(User, UserAdmin)
<commit_before>from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) <commit_msg>Add User back into admin<commit_after>from django.contrib.admin.sites import AdminSite from django.conf.urls.defaults import patterns, url from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet class ShellsAdmin(AdminSite): def get_urls(self): urls = super(ShellsAdmin, self).get_urls() my_urls = patterns('', url('upload_images/', self.admin_view(ShellsImagesUploader.as_view()), name="upload-images"), url(r'^upload/', self.admin_view(upload_species_spreadsheet), name='upload_species_spreadsheet'), ) return my_urls + urls shells_admin = ShellsAdmin() from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin from shells.models import Species, Specimen, SpeciesRepresentation shells_admin.register(Species, SpeciesAdmin) shells_admin.register(Specimen, SpecimenAdmin) shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin) from django.contrib.auth.admin import UserAdmin from django.contrib.auth.models import User shells_admin.register(User, UserAdmin)
52d38e360b14fcfad01f87ff1e9ca5db27004877
src/comms/admin.py
src/comms/admin.py
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions") list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage') list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
Python
bsd-3-clause
ypwalter/evennia,TheTypoMaster/evennia,TheTypoMaster/evennia,mrkulk/text-world,mrkulk/text-world,titeuf87/evennia,ergodicbreak/evennia,mrkulk/text-world,feend78/evennia,shollen/evennia,jamesbeebop/evennia,shollen/evennia,feend78/evennia,ergodicbreak/evennia,feend78/evennia,titeuf87/evennia,mrkulk/text-world,jamesbeebop/evennia,emergebtc/evennia,ypwalter/evennia,titeuf87/evennia,TheTypoMaster/evennia,emergebtc/evennia,titeuf87/evennia,ypwalter/evennia,feend78/evennia,emergebtc/evennia,jamesbeebop/evennia,ergodicbreak/evennia
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions") list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin) Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage') list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
<commit_before># # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions") list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin) <commit_msg>Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.<commit_after>
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage') list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions") list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin) Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.# # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage') list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
<commit_before># # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions") list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin) <commit_msg>Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.<commit_after># # This sets up how models are displayed # in the web admin interface. # from django.contrib import admin from src.comms.models import ChannelDB class MsgAdmin(admin.ModelAdmin): list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers', 'db_channels', 'db_message', 'db_lock_storage') list_display_links = ("id",) ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels'] #readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels'] search_fields = ['id', '^db_date_sent', '^db_message'] save_as = True save_on_top = True list_select_related = True #admin.site.register(Msg, MsgAdmin) class ChannelAdmin(admin.ModelAdmin): list_display = ('id', 'db_key', 'db_lock_storage') list_display_links = ("id", 'db_key') ordering = ["db_key"] search_fields = ['id', 'db_key', 'db_aliases'] save_as = True save_on_top = True list_select_related = True fieldsets = ( (None, {'fields': (('db_key',), 'db_lock_storage')}), ) admin.site.register(ChannelDB, ChannelAdmin)
21b6a5573190848b93de930b9d41e1ac766c18bc
src/epiweb/urls.py
src/epiweb/urls.py
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), )
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^accounts/', include('registration.backends.default.urls')), )
Add user registration URLs. Use what django-registration provides for the moment.
Add user registration URLs. Use what django-registration provides for the moment.
Python
agpl-3.0
ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), ) Add user registration URLs. Use what django-registration provides for the moment.
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^accounts/', include('registration.backends.default.urls')), )
<commit_before>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), ) <commit_msg>Add user registration URLs. Use what django-registration provides for the moment.<commit_after>
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^accounts/', include('registration.backends.default.urls')), )
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), ) Add user registration URLs. Use what django-registration provides for the moment.from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^accounts/', include('registration.backends.default.urls')), )
<commit_before>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), ) <commit_msg>Add user registration URLs. Use what django-registration provides for the moment.<commit_after>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() import epiweb.apps.survey.urls urlpatterns = patterns('', # Example: # (r'^epiweb/', include('epiweb.foo.urls')), # Uncomment the admin/doc line below and add 'django.contrib.admindocs' # to INSTALLED_APPS to enable admin documentation: # (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^accounts/', include('registration.backends.default.urls')), )
017de01e8a1ec8f49069cf546e89652b4ddb8e39
tests/test_create_template.py
tests/test_create_template.py
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess @pytest.fixture def output_dir(tmpdir): return str(tmpdir.mkdir('output')) def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
Remove output_dir fixture from test
Remove output_dir fixture from test
Python
mit
pytest-dev/cookiecutter-pytest-plugin
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess @pytest.fixture def output_dir(tmpdir): return str(tmpdir.mkdir('output')) def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project)) Remove output_dir fixture from test
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
<commit_before># -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess @pytest.fixture def output_dir(tmpdir): return str(tmpdir.mkdir('output')) def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project)) <commit_msg>Remove output_dir fixture from test<commit_after>
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess @pytest.fixture def output_dir(tmpdir): return str(tmpdir.mkdir('output')) def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project)) Remove output_dir fixture from test# -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
<commit_before># -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess @pytest.fixture def output_dir(tmpdir): return str(tmpdir.mkdir('output')) def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project)) <commit_msg>Remove output_dir fixture from test<commit_after># -*- coding: utf-8 -*- """ test_create_template -------------------- """ import os import pytest import subprocess def run_tox(plugin): """Run the tox suite of the newly created plugin.""" try: subprocess.check_call([ 'tox', plugin, '-c', os.path.join(plugin, 'tox.ini'), '-e', 'py' ]) except subprocess.CalledProcessError as e: pytest.fail(e) def test_run_cookiecutter_and_plugin_tests(cookies): """Create a new plugin via cookiecutter and run its tests.""" result = cookies.bake() assert result.project.basename == 'pytest-foobar' assert result.project.isdir() run_tox(str(result.project))
a6e9438cd7aa205fb1ae1f20d192b9139bcd7f95
thecure/game.py
thecure/game.py
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((800, 600)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((1024, 768)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
Increase the resolution to 1024x768.
Increase the resolution to 1024x768. This will make it easier to see things like the house. 800x600 is just too small.
Python
mit
chipx86/the-cure
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((800, 600)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit() Increase the resolution to 1024x768. This will make it easier to see things like the house. 800x600 is just too small.
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((1024, 768)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
<commit_before>import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((800, 600)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit() <commit_msg>Increase the resolution to 1024x768. This will make it easier to see things like the house. 800x600 is just too small.<commit_after>
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((1024, 768)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((800, 600)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit() Increase the resolution to 1024x768. This will make it easier to see things like the house. 800x600 is just too small.import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((1024, 768)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
<commit_before>import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((800, 600)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit() <commit_msg>Increase the resolution to 1024x768. This will make it easier to see things like the house. 800x600 is just too small.<commit_after>import pygame from pygame.locals import * from thecure.engine import TheCureEngine def main(): pygame.init() version = pygame.__version__.split('.') if int(version[0]) <= 1 and int(version[1]) < 9: print 'This game requires pygame 1.9 or higher.' return screen = pygame.display.set_mode((1024, 768)) pygame.display.set_caption('The Cure') engine = TheCureEngine(screen) engine.run() pygame.quit()
18b0ddbbca429778a70f1e9b7f7d5140eb88d68f
tests/test_fs.py
tests/test_fs.py
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c")
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") def test_cmp(): assert Path("/a/b") < Path("/a/c") assert Path("/a/c") > Path("/a/b") assert Path("/a/2") < Path("/b/1")
Add test to Path compare.
Add test to Path compare.
Python
mit
andrewguy9/farmfs,andrewguy9/farmfs
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") Add test to Path compare.
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") def test_cmp(): assert Path("/a/b") < Path("/a/c") assert Path("/a/c") > Path("/a/b") assert Path("/a/2") < Path("/b/1")
<commit_before>from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") <commit_msg>Add test to Path compare.<commit_after>
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") def test_cmp(): assert Path("/a/b") < Path("/a/c") assert Path("/a/c") > Path("/a/b") assert Path("/a/2") < Path("/b/1")
from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") Add test to Path compare.from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") def test_cmp(): assert Path("/a/b") < Path("/a/c") assert Path("/a/c") > Path("/a/b") assert Path("/a/2") < Path("/b/1")
<commit_before>from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") <commit_msg>Add test to Path compare.<commit_after>from farmfs.fs import normpath as _normalize from farmfs.fs import userPath2Path as up2p from farmfs.fs import Path def test_normalize_abs(): assert _normalize("/") == "/" assert _normalize("/a") == "/a" assert _normalize("/a/") == "/a" assert _normalize("/a/b") == "/a/b" assert _normalize("/a/b/") == "/a/b" assert _normalize("/a//b") == "/a/b" assert _normalize("/a//b//") == "/a/b" def test_normalize_relative(): assert _normalize("a") == "a" assert _normalize("a/") == "a" assert _normalize("a/b") == "a/b" assert _normalize("a/b/") == "a/b" assert _normalize("a//b") == "a/b" assert _normalize("a//b//") == "a/b" def test_userPath2Path(): assert up2p("c", Path("/a/b")) == Path("/a/b/c") assert up2p("/c", Path("/a/b")) == Path("/c") def test_cmp(): assert Path("/a/b") < Path("/a/c") assert Path("/a/c") > Path("/a/b") assert Path("/a/2") < Path("/b/1")
c41d0a9f03e66fdc20fb093aaad87cdd6f60461e
studies/helpers.py
studies/helpers.py
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.settings import EMAIL_FROM_ADDRESS, BASE_URL # TODO: celery taskify def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
Add decorator to send_mail function to celery taskify.
Add decorator to send_mail function to celery taskify.
Python
apache-2.0
pattisdr/lookit-api,pattisdr/lookit-api,CenterForOpenScience/lookit-api,CenterForOpenScience/lookit-api,pattisdr/lookit-api,CenterForOpenScience/lookit-api
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.settings import EMAIL_FROM_ADDRESS, BASE_URL # TODO: celery taskify def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() Add decorator to send_mail function to celery taskify.
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
<commit_before>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.settings import EMAIL_FROM_ADDRESS, BASE_URL # TODO: celery taskify def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() <commit_msg>Add decorator to send_mail function to celery taskify.<commit_after>
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.settings import EMAIL_FROM_ADDRESS, BASE_URL # TODO: celery taskify def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() Add decorator to send_mail function to celery taskify.from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
<commit_before>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.settings import EMAIL_FROM_ADDRESS, BASE_URL # TODO: celery taskify def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send() <commit_msg>Add decorator to send_mail function to celery taskify.<commit_after>from django.core.mail.message import EmailMultiAlternatives from django.template.loader import get_template from project.celery import app from project.settings import EMAIL_FROM_ADDRESS, BASE_URL @app.task def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context): """ Helper for sending templated email :param str template_name: Name of the template to send. There should exist a txt and html version :param str subject: Subject line of the email :param str from_address: From address for email :param list to_addresses: List of addresses to email. If str is provided, wrapped in list :param list cc: List of addresses to carbon copy :param list bcc: List of addresses to blind carbon copy :param str custom_message Custom email message - for use instead of a template :kwargs: Context vars for the email template """ context['base_url'] = BASE_URL text_content = get_template('emails/{}.txt'.format(template_name)).render(context) html_content = get_template('emails/{}.html'.format(template_name)).render(context) if not isinstance(to_addresses, list): to_addresses = [to_addresses] from_address = from_email or EMAIL_FROM_ADDRESS email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc) email.attach_alternative(html_content, 'text/html') email.send()
49a1548399fa822515920d910ec6ea6a6c813bca
threadpool.py
threadpool.py
from __future__ import with_statement import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time @threado.stream def run(inner, self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() result = yield channel inner.finish(result) def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
from __future__ import with_statement import sys import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time def run(self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() return channel def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
Python
mit
abusesa/idiokit
from __future__ import with_statement import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time @threado.stream def run(inner, self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() result = yield channel inner.finish(result) def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
from __future__ import with_statement import sys import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time def run(self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() return channel def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
<commit_before>from __future__ import with_statement import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time @threado.stream def run(inner, self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() result = yield channel inner.finish(result) def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run <commit_msg>Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.<commit_after>
from __future__ import with_statement import sys import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time def run(self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() return channel def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
from __future__ import with_statement import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time @threado.stream def run(inner, self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() result = yield channel inner.finish(result) def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.from __future__ import with_statement import sys import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time def run(self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() return channel def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
<commit_before>from __future__ import with_statement import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time @threado.stream def run(inner, self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() result = yield channel inner.finish(result) def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run <commit_msg>Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.<commit_after>from __future__ import with_statement import sys import threado import threading import Queue class ThreadPool(object): def __init__(self, idle_time=5.0): self.lock = threading.Lock() self.threads = list() self.idle_time = idle_time def run(self, func, *args, **keys): with self.lock: if self.threads: thread, queue = self.threads.pop() else: queue = Queue.Queue() thread = threading.Thread(target=self._thread, args=(queue,)) thread.setDaemon(True) channel = threado.Channel() queue.put((channel, func, args, keys)) if not thread.isAlive(): thread.start() return channel def _thread(self, queue): item = threading.currentThread(), queue while True: try: task = queue.get(True, self.idle_time) except Queue.Empty: with self.lock: if item not in self.threads: continue self.threads.remove(item) return if task is None: return channel, func, args, keys = task try: result = func(*args, **keys) except: self.threads.append(item) channel.rethrow() else: self.threads.append(item) channel.finish(result) thread_pool = ThreadPool() run = thread_pool.run
b252592eb40263994317d88ced43ddc4669a4975
tests/run_tests.py
tests/run_tests.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: main(args)
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) return unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
Fix return code when running unittests.
Fix return code when running unittests.
Python
bsd-3-clause
AndiDog/sqlparse,actsasgeek/sqlparse,zhongdai/sqlparse,tailhook/sqlparse,AndiDog/sqlparse,Yelp/sqlparse,Yelp/sqlparse,benekastah/sqlparse,actsasgeek/sqlparse,tailhook/sqlparse,Yelp/sqlparse,MariaPet/sqlparse,payne/sqlparse,adamgreenhall/sqlparse,andialbrecht/sqlparse,tailhook/sqlparse,adamgreenhall/sqlparse,payne/sqlparse,adamgreenhall/sqlparse,payne/sqlparse,MariaPet/sqlparse,zhongdai/sqlparse,vmuriart/sqlparse,AndiDog/sqlparse,zhongdai/sqlparse,actsasgeek/sqlparse
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: main(args) Fix return code when running unittests.
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) return unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: main(args) <commit_msg>Fix return code when running unittests.<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) return unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: main(args) Fix return code when running unittests.#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) return unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: main(args) <commit_msg>Fix return code when running unittests.<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """Test runner for sqlparse.""" import optparse import os import sys import unittest test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../')) if test_mod not in sys.path: sys.path.insert(1, test_mod) parser = optparse.OptionParser() parser.add_option('-P', '--profile', help='Create hotshot profile.', action='store_true', default=False) def main(args): """Create a TestSuite and run it.""" loader = unittest.TestLoader() suite = unittest.TestSuite() fnames = [os.path.split(f)[-1] for f in args] for fname in os.listdir(os.path.dirname(__file__)): if (not fname.startswith('test_') or not fname.endswith('.py') or (fnames and fname not in fnames)): continue modname = os.path.splitext(fname)[0] mod = __import__(os.path.splitext(fname)[0]) suite.addTests(loader.loadTestsFromModule(mod)) return unittest.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': opts, args = parser.parse_args() if opts.profile: import hotshot prof = hotshot.Profile("sqlparse.prof") prof.runcall(main, args) prof.close() else: result = main(args) if not result.wasSuccessful(): return_code = 1 else: return_code = 0 sys.exit(return_code)
12ec1cf9084789b9e2022eb0d1d55b553db06cb5
tests/test_util.py
tests/test_util.py
import util from nose.tools import assert_equal class TestPick(): def check(self, filenames, expected, k, randomized): result = util.pick(filenames, k, randomized) assert_equal(result, expected) def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] self.check(filenames, expected, k=None, randomized=False)
import util from nose.tools import assert_equal, assert_true, raises class TestPick(): def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] result = util.pick(filenames, randomized=False) assert_equal(result, expected) def test_k_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt'] result = util.pick(filenames, k=3, randomized=False) assert_equal(result, expected) def test_all_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames) assert_equal(sorted(filenames), sorted(result)) def test_k_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames, k=3) for r in result: assert_true(r in filenames) @raises(ValueError) def test_negative_k(self): util.pick([], k=-2)
Fix unit test for util.py
Fix unit test for util.py
Python
mit
kemskems/otdet
import util from nose.tools import assert_equal class TestPick(): def check(self, filenames, expected, k, randomized): result = util.pick(filenames, k, randomized) assert_equal(result, expected) def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] self.check(filenames, expected, k=None, randomized=False) Fix unit test for util.py
import util from nose.tools import assert_equal, assert_true, raises class TestPick(): def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] result = util.pick(filenames, randomized=False) assert_equal(result, expected) def test_k_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt'] result = util.pick(filenames, k=3, randomized=False) assert_equal(result, expected) def test_all_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames) assert_equal(sorted(filenames), sorted(result)) def test_k_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames, k=3) for r in result: assert_true(r in filenames) @raises(ValueError) def test_negative_k(self): util.pick([], k=-2)
<commit_before>import util from nose.tools import assert_equal class TestPick(): def check(self, filenames, expected, k, randomized): result = util.pick(filenames, k, randomized) assert_equal(result, expected) def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] self.check(filenames, expected, k=None, randomized=False) <commit_msg>Fix unit test for util.py<commit_after>
import util from nose.tools import assert_equal, assert_true, raises class TestPick(): def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] result = util.pick(filenames, randomized=False) assert_equal(result, expected) def test_k_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt'] result = util.pick(filenames, k=3, randomized=False) assert_equal(result, expected) def test_all_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames) assert_equal(sorted(filenames), sorted(result)) def test_k_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames, k=3) for r in result: assert_true(r in filenames) @raises(ValueError) def test_negative_k(self): util.pick([], k=-2)
import util from nose.tools import assert_equal class TestPick(): def check(self, filenames, expected, k, randomized): result = util.pick(filenames, k, randomized) assert_equal(result, expected) def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] self.check(filenames, expected, k=None, randomized=False) Fix unit test for util.pyimport util from nose.tools import assert_equal, assert_true, raises class TestPick(): def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] result = util.pick(filenames, randomized=False) assert_equal(result, expected) def test_k_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt'] result = util.pick(filenames, k=3, randomized=False) assert_equal(result, expected) def test_all_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames) assert_equal(sorted(filenames), sorted(result)) def test_k_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames, k=3) for r in result: assert_true(r in filenames) @raises(ValueError) def test_negative_k(self): util.pick([], k=-2)
<commit_before>import util from nose.tools import assert_equal class TestPick(): def check(self, filenames, expected, k, randomized): result = util.pick(filenames, k, randomized) assert_equal(result, expected) def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] self.check(filenames, expected, k=None, randomized=False) <commit_msg>Fix unit test for util.py<commit_after>import util from nose.tools import assert_equal, assert_true, raises class TestPick(): def test_all_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt'] result = util.pick(filenames, randomized=False) assert_equal(result, expected) def test_k_sequential(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] expected = ['e-0.txt', 'd-1.txt', 'b-2.txt'] result = util.pick(filenames, k=3, randomized=False) assert_equal(result, expected) def test_all_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames) assert_equal(sorted(filenames), sorted(result)) def test_k_random(self): filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt'] result = util.pick(filenames, k=3) for r in result: assert_true(r in filenames) @raises(ValueError) def test_negative_k(self): util.pick([], k=-2)
048107baf565c26a21421ef8f703be5cf7591cee
src/ggrc/fulltext/__init__.py
src/ggrc/fulltext/__init__.py
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags, **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags="", **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
Change tags to kwarg in full text record
Change tags to kwarg in full text record This change will allow us to set the tags value with kwargs that are set in custom attribute values.
Python
apache-2.0
edofic/ggrc-core,AleksNeStu/ggrc-core,josthkko/ggrc-core,andrei-karalionak/ggrc-core,AleksNeStu/ggrc-core,andrei-karalionak/ggrc-core,AleksNeStu/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core,VinnieJohns/ggrc-core,edofic/ggrc-core,VinnieJohns/ggrc-core,j0gurt/ggrc-core,plamut/ggrc-core,edofic/ggrc-core,j0gurt/ggrc-core,plamut/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,selahssea/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,AleksNeStu/ggrc-core,josthkko/ggrc-core,plamut/ggrc-core,andrei-karalionak/ggrc-core,selahssea/ggrc-core,kr41/ggrc-core,selahssea/ggrc-core,edofic/ggrc-core,josthkko/ggrc-core,kr41/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,VinnieJohns/ggrc-core
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags, **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer) Change tags to kwarg in full text record This change will allow us to set the tags value with kwargs that are set in custom attribute values.
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags="", **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
<commit_before># Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags, **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer) <commit_msg>Change tags to kwarg in full text record This change will allow us to set the tags value with kwargs that are set in custom attribute values.<commit_after>
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags="", **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags, **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer) Change tags to kwarg in full text record This change will allow us to set the tags value with kwargs that are set in custom attribute values.# Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags="", **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
<commit_before># Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags, **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer) <commit_msg>Change tags to kwarg in full text record This change will allow us to set the tags value with kwargs that are set in custom attribute values.<commit_after># Copyright (C) 2016 Google Inc. # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> from ggrc.extensions import get_extension_instance class Indexer(object): def __init__(self, settings): pass def create_record(self, record): raise NotImplementedError() def update_record(self, record): raise NotImplementedError() def delete_record(self, key): raise NotImplementedError() def search(self, terms): raise NotImplementedError() class Record(object): def __init__(self, key, type, context_id, tags="", **kwargs): self.key = key self.type = type self.context_id = context_id self.tags = tags self.properties = kwargs def resolve_default_text_indexer(): from ggrc import settings db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0] return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme) def get_indexer(indexer=[]): return get_extension_instance( 'FULLTEXT_INDEXER', resolve_default_text_indexer)
f6bcc5af968c5c8b81fd6db10bad50b1fadcb5d1
discover_road_runner/meep_meep/tests/test_meep_meep.py
discover_road_runner/meep_meep/tests/test_meep_meep.py
import unittest from django.test import SimpleTestCase class BasicTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
import unittest from django.test import SimpleTestCase class CombinedTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
Rename test as it really is a full integration test.
Rename test as it really is a full integration test.
Python
mit
pzrq/discover-road-runner
import unittest from django.test import SimpleTestCase class BasicTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!') Rename test as it really is a full integration test.
import unittest from django.test import SimpleTestCase class CombinedTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
<commit_before>import unittest from django.test import SimpleTestCase class BasicTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!') <commit_msg>Rename test as it really is a full integration test.<commit_after>
import unittest from django.test import SimpleTestCase class CombinedTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
import unittest from django.test import SimpleTestCase class BasicTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!') Rename test as it really is a full integration test.import unittest from django.test import SimpleTestCase class CombinedTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
<commit_before>import unittest from django.test import SimpleTestCase class BasicTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!') <commit_msg>Rename test as it really is a full integration test.<commit_after>import unittest from django.test import SimpleTestCase class CombinedTest(SimpleTestCase): def test_success(self): self.assertEqual(1 + 1, 2) def test_error(self): raise ValueError def test_fail(self): self.fail() def test_skip(self): self.skipTest('Fix me later?') @unittest.expectedFailure def test_decorator_expected_fail(self): self.fail() @unittest.skip def test_decorator_skip(self): print('Never get here!')
ff9a4f89d81c2059f51f7346700ce16972f04e36
django_basic_tinymce_flatpages/admin.py
django_basic_tinymce_flatpages/admin.py
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', default={'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', {'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
Fix getattr() takes no keyword arguments
Fix getattr() takes no keyword arguments
Python
bsd-3-clause
ad-m/django-basic-tinymce-flatpages
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', default={'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin) Fix getattr() takes no keyword arguments
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', {'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
<commit_before>from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', default={'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin) <commit_msg>Fix getattr() takes no keyword arguments<commit_after>
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', {'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', default={'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin) Fix getattr() takes no keyword argumentsfrom django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', {'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
<commit_before>from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', default={'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin) <commit_msg>Fix getattr() takes no keyword arguments<commit_after>from django.conf import settings from django.contrib import admin from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin from django.contrib.flatpages.models import FlatPage from django.utils.module_loading import import_string FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE') FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS', {'attrs': {'cols': 100, 'rows': 15}}) class PageForm(FlatpageForm): class Meta: model = FlatPage widgets = { 'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS), } class PageAdmin(FlatPageAdmin): """ Page Admin """ form = PageForm admin.site.unregister(FlatPage) admin.site.register(FlatPage, PageAdmin)
0bf00b40e84a5c5fbcdbeb7b81911998e3f1081a
src/idea/tests/smoke_tests.py
src/idea/tests/smoke_tests.py
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="test1@example.com", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "test1@example.com" user.first_name = 'first' user.last_name = 'last' user.email = '"test1@example.com"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="test1@example.com") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
Use fixtures for smoke tests
Use fixtures for smoke tests
Python
cc0-1.0
cmc333333/idea-box,m3brown/idea-box,18F/idea-box,cmc333333/idea-box,CapeSepias/idea-box,geomapdev/idea-box,CapeSepias/idea-box,cmc333333/idea-box,18F/idea-box,geomapdev/idea-box,geomapdev/idea-box,18F/idea-box,CapeSepias/idea-box,m3brown/idea-box
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="test1@example.com", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "test1@example.com" user.first_name = 'first' user.last_name = 'last' user.email = '"test1@example.com"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code) Use fixtures for smoke tests
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="test1@example.com") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
<commit_before>import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="test1@example.com", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "test1@example.com" user.first_name = 'first' user.last_name = 'last' user.email = '"test1@example.com"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code) <commit_msg>Use fixtures for smoke tests<commit_after>
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="test1@example.com") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="test1@example.com", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "test1@example.com" user.first_name = 'first' user.last_name = 'last' user.email = '"test1@example.com"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code) Use fixtures for smoke testsimport os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="test1@example.com") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
<commit_before>import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state'] @fixture def user(self): try: from collab.django_factories import UserF return UserF(username="test1@example.com", person__title='') except ImportError: from django.contrib.auth.models import User user = User() user.username = "test1@example.com" user.first_name = 'first' user.last_name = 'last' user.email = '"test1@example.com"' user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs=' user.is_staff = False user.is_active = True user.is_superuser = False user.last_login = timezone.now() user.date_joined = timezone.now() user.save() return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code) <commit_msg>Use fixtures for smoke tests<commit_after>import os from django.utils import timezone from django_webtest import WebTest from exam.decorators import fixture from exam.cases import Exam from django.core.urlresolvers import reverse from django.contrib.auth.models import User class SmokeTest(Exam, WebTest): csrf_checks = False fixtures = ['state', 'core-test-fixtures'] @fixture def user(self): user = User.objects.get(username="test1@example.com") return user def get(self, url): return self.app.get(url, user=self.user) def test_idea_home(self): page = self.get(reverse('idea:idea_list')) self.assertEquals(200, page.status_code)
b77a3f47876d824d2e0f1c009a6d580fc5d41ec6
accelerator/migrations/0019_add_deferred_user_role.py
accelerator/migrations/0019_add_deferred_user_role.py
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') user_role = UserRole.objects.create( name=DEFERRED_MENTOR, sort_order='17') for program in Program.objects.all(): name = "{} {} Deferred Mentor".format( program.start_date.year, program.program_family.name) return ProgramRole.objects.get_or_create( program=program, user_role=user_role, defaults={'name': name}) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') if UserRole.objects.filter(name=DEFERRED_MENTOR).exists(): user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0] else: user_role = UserRole.objects.create(name=DEFERRED_MENTOR, sort_order=17) for program in Program.objects.all(): if not ProgramRole.objects.filter(user_role=user_role, program=program).exists(): name = "{} {} ({}-{})".format( (program.end_date.year if program.end_date else ""), DEFERRED_MENTOR, program.program_family.url_slug.upper(), program.pk) ProgramRole.objects.get_or_create( program=program, user_role=user_role, name=name) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
Make changes to the migration file
[AC-7594] Make changes to the migration file
Python
mit
masschallenge/django-accelerator,masschallenge/django-accelerator
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') user_role = UserRole.objects.create( name=DEFERRED_MENTOR, sort_order='17') for program in Program.objects.all(): name = "{} {} Deferred Mentor".format( program.start_date.year, program.program_family.name) return ProgramRole.objects.get_or_create( program=program, user_role=user_role, defaults={'name': name}) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ] [AC-7594] Make changes to the migration file
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') if UserRole.objects.filter(name=DEFERRED_MENTOR).exists(): user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0] else: user_role = UserRole.objects.create(name=DEFERRED_MENTOR, sort_order=17) for program in Program.objects.all(): if not ProgramRole.objects.filter(user_role=user_role, program=program).exists(): name = "{} {} ({}-{})".format( (program.end_date.year if program.end_date else ""), DEFERRED_MENTOR, program.program_family.url_slug.upper(), program.pk) ProgramRole.objects.get_or_create( program=program, user_role=user_role, name=name) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
<commit_before># Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') user_role = UserRole.objects.create( name=DEFERRED_MENTOR, sort_order='17') for program in Program.objects.all(): name = "{} {} Deferred Mentor".format( program.start_date.year, program.program_family.name) return ProgramRole.objects.get_or_create( program=program, user_role=user_role, defaults={'name': name}) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ] <commit_msg>[AC-7594] Make changes to the migration file<commit_after>
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') if UserRole.objects.filter(name=DEFERRED_MENTOR).exists(): user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0] else: user_role = UserRole.objects.create(name=DEFERRED_MENTOR, sort_order=17) for program in Program.objects.all(): if not ProgramRole.objects.filter(user_role=user_role, program=program).exists(): name = "{} {} ({}-{})".format( (program.end_date.year if program.end_date else ""), DEFERRED_MENTOR, program.program_family.url_slug.upper(), program.pk) ProgramRole.objects.get_or_create( program=program, user_role=user_role, name=name) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') user_role = UserRole.objects.create( name=DEFERRED_MENTOR, sort_order='17') for program in Program.objects.all(): name = "{} {} Deferred Mentor".format( program.start_date.year, program.program_family.name) return ProgramRole.objects.get_or_create( program=program, user_role=user_role, defaults={'name': name}) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ] [AC-7594] Make changes to the migration file# Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') if UserRole.objects.filter(name=DEFERRED_MENTOR).exists(): user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0] else: user_role = UserRole.objects.create(name=DEFERRED_MENTOR, sort_order=17) for program in Program.objects.all(): if not ProgramRole.objects.filter(user_role=user_role, program=program).exists(): name = "{} {} ({}-{})".format( (program.end_date.year if program.end_date else ""), DEFERRED_MENTOR, program.program_family.url_slug.upper(), program.pk) ProgramRole.objects.get_or_create( program=program, user_role=user_role, name=name) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
<commit_before># Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') user_role = UserRole.objects.create( name=DEFERRED_MENTOR, sort_order='17') for program in Program.objects.all(): name = "{} {} Deferred Mentor".format( program.start_date.year, program.program_family.name) return ProgramRole.objects.get_or_create( program=program, user_role=user_role, defaults={'name': name}) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ] <commit_msg>[AC-7594] Make changes to the migration file<commit_after># Generated by Django 2.2.10 on 2020-04-09 21:24 from django.db import migrations def add_deferred_user_role(apps, schema_editor): DEFERRED_MENTOR = 'Deferred Mentor' UserRole = apps.get_model('accelerator', 'UserRole') Program = apps.get_model('accelerator', 'Program') ProgramRole = apps.get_model('accelerator', 'ProgramRole') if UserRole.objects.filter(name=DEFERRED_MENTOR).exists(): user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0] else: user_role = UserRole.objects.create(name=DEFERRED_MENTOR, sort_order=17) for program in Program.objects.all(): if not ProgramRole.objects.filter(user_role=user_role, program=program).exists(): name = "{} {} ({}-{})".format( (program.end_date.year if program.end_date else ""), DEFERRED_MENTOR, program.program_family.url_slug.upper(), program.pk) ProgramRole.objects.get_or_create( program=program, user_role=user_role, name=name) class Migration(migrations.Migration): dependencies = [ ('accelerator', '0018_make_location_nonrequired'), ] operations = [ migrations.RunPython(add_deferred_user_role, migrations.RunPython.noop) ]
4574d25ade5c18d6c15ac6d427f4fbd4cb2f0f04
braid/info.py
braid/info.py
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ('rhel', '/etc/redhat-release'), ('debian', '/etc/debian_version'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
Add debian-squeeze support for os detection.
Add debian-squeeze support for os detection.
Python
mit
alex/braid,alex/braid
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch'] Add debian-squeeze support for os detection.
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ('rhel', '/etc/redhat-release'), ('debian', '/etc/debian_version'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
<commit_before>from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch'] <commit_msg>Add debian-squeeze support for os detection.<commit_after>
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ('rhel', '/etc/redhat-release'), ('debian', '/etc/debian_version'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch'] Add debian-squeeze support for os detection.from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ('rhel', '/etc/redhat-release'), ('debian', '/etc/debian_version'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
<commit_before>from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch'] <commit_msg>Add debian-squeeze support for os detection.<commit_after>from fabric.api import run, quiet from braid import succeeds, cacheInEnvironment @cacheInEnvironment def distroName(): """ Get the name of the distro. """ with quiet(): lsb = run('/usr/bin/lsb_release --id --short', warn_only=True) if lsb.succeeded: return lsb.lower() distros = [ ('centos', '/etc/centos-release'), ('fedora', '/etc/fedora-release'), ('rhel', '/etc/redhat-release'), ('debian', '/etc/debian_version'), ] for distro, sentinel in distros: if succeeds('/usr/bin/test -f {}'.format(sentinel)): return distro def distroFamily(): """ Get the family of the distro. @returns: C{'debian'} or C{'fedora'} """ families = { 'debian': ['debian', 'ubuntu'], 'fedora': ['fedora', 'centos', 'rhel'], } distro = distroName() for family, members in families.iteritems(): if distro in members: return family return 'other' @cacheInEnvironment def arch(): """ Get the architechture of the machine. """ return run('/bin/uname --machine') @cacheInEnvironment def isRoot(): """ Check if the current user is root. """ return run('id -u') == '0' __all__ = ['distroName', 'distroFamily', 'arch']
ee7ced467a7b87e71aa5a1df4c828e672d0b9870
Utils/py/BallDetection/RegressionNetwork/evaluate.py
Utils/py/BallDetection/RegressionNetwork/evaluate.py
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") print("loss: {} precision: {}".format(result[0], result[1]))
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") for idx in range(0, len(result)): print(model.metrics_names[idx] + ":", result[idx])
Print all evaluation metrics associated with the model
Print all evaluation metrics associated with the model
Python
apache-2.0
BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") print("loss: {} precision: {}".format(result[0], result[1])) Print all evaluation metrics associated with the model
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") for idx in range(0, len(result)): print(model.metrics_names[idx] + ":", result[idx])
<commit_before>#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") print("loss: {} precision: {}".format(result[0], result[1])) <commit_msg>Print all evaluation metrics associated with the model<commit_after>
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") for idx in range(0, len(result)): print(model.metrics_names[idx] + ":", result[idx])
#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") print("loss: {} precision: {}".format(result[0], result[1])) Print all evaluation metrics associated with the model#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") for idx in range(0, len(result)): print(model.metrics_names[idx] + ":", result[idx])
<commit_before>#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") print("loss: {} precision: {}".format(result[0], result[1])) <commit_msg>Print all evaluation metrics associated with the model<commit_after>#!/usr/bin/env python3 import argparse import pickle import tensorflow.keras as keras import numpy as np import sys import cv2 parser = argparse.ArgumentParser(description='Train the network given ') parser.add_argument('-b', '--database-path', dest='imgdb_path', help='Path to the image database containing test data.' 'Default is img.db in current folder.') parser.add_argument('-m', '--model-path', dest='model_path', help='Store the trained model using this path. Default is model.h5.') args = parser.parse_args() imgdb_path = "img.db" model_path = "model.h5" res = {"x": 16, "y": 16} if args.model_path is not None: model_path = args.model_path if args.imgdb_path is not None: imgdb_path = args.imgdb_path with open(imgdb_path, "rb") as f: mean = pickle.load(f) print("mean=" + str(mean)) x = pickle.load(f) y = pickle.load(f) model = keras.models.load_model(model_path) print(model.summary()) x = np.array(x) y = np.array(y) result = model.evaluate(x,y) print("Evaluation result") print("=================") for idx in range(0, len(result)): print(model.metrics_names[idx] + ":", result[idx])
d25603818e6af0b99ee1a6add0a7e182037d7a12
tests/test_set_pref.py
tests/test_set_pref.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs() as recent_refs: languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs(): yvs.shared.update_recent_refs(['8/mat.5']) languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) nose.assert_equal(len(yvs.shared.get_recent_refs()), 0) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
Verify that recent list is cleared when language is changed
Verify that recent list is cleared when language is changed
Python
mit
caleb531/youversion-suggest,caleb531/youversion-suggest
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs() as recent_refs: languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version) Verify that recent list is cleared when language is changed
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs(): yvs.shared.update_recent_refs(['8/mat.5']) languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) nose.assert_equal(len(yvs.shared.get_recent_refs()), 0) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs() as recent_refs: languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version) <commit_msg>Verify that recent list is cleared when language is changed<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs(): yvs.shared.update_recent_refs(['8/mat.5']) languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) nose.assert_equal(len(yvs.shared.get_recent_refs()), 0) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs() as recent_refs: languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version) Verify that recent list is cleared when language is changed#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs(): yvs.shared.update_recent_refs(['8/mat.5']) languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) nose.assert_equal(len(yvs.shared.get_recent_refs()), 0) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs() as recent_refs: languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) # Check if new values have been saved to preferences prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version) <commit_msg>Verify that recent list is cleared when language is changed<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- from __future__ import unicode_literals import nose.tools as nose import yv_suggest.set_pref as yvs import context_managers as ctx def test_set_language(): """should set preferred language""" with ctx.preserve_prefs() as prefs: with ctx.preserve_recent_refs(): yvs.shared.update_recent_refs(['8/mat.5']) languages = yvs.shared.get_languages() for language in languages: if language['id'] != prefs['language']: new_language = language['id'] break yvs.main('language:{}'.format(new_language)) prefs = yvs.shared.get_prefs() bible = yvs.shared.get_bible_data(prefs['language']) nose.assert_equal(prefs['language'], new_language) nose.assert_equal(prefs['version'], bible['default_version']) nose.assert_equal(len(yvs.shared.get_recent_refs()), 0) def test_set_version(): """should set preferred version""" with ctx.preserve_prefs() as prefs: bible = yvs.shared.get_bible_data(prefs['language']) versions = bible['versions'] for version in versions: if version['id'] != prefs['language']: new_version = version['id'] break yvs.main('version:{}'.format(new_version)) prefs = yvs.shared.get_prefs() nose.assert_equal(prefs['version'], new_version)
88098475358aaee18d32a1ad2c4a4301672bca0e
account_move_fiscal_month/models/account_move_line.py
account_move_fiscal_month/models/account_move_line.py
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', store=True, copy=False)
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', )
Remove unneeded and inefficient "store=True"
[FIX] Remove unneeded and inefficient "store=True"
Python
agpl-3.0
OCA/account-financial-tools,OCA/account-financial-tools
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', store=True, copy=False) [FIX] Remove unneeded and inefficient "store=True"
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', )
<commit_before># Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', store=True, copy=False) <commit_msg>[FIX] Remove unneeded and inefficient "store=True"<commit_after>
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', )
# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', store=True, copy=False) [FIX] Remove unneeded and inefficient "store=True"# Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', )
<commit_before># Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', store=True, copy=False) <commit_msg>[FIX] Remove unneeded and inefficient "store=True"<commit_after># Copyright 2017 ACSONE SA/NV # License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl). from odoo import fields, models class AccountMoveLine(models.Model): _inherit = 'account.move.line' date_range_fm_id = fields.Many2one( related='move_id.date_range_fm_id', )
68a7fd8a444a8c568d716db11849f58ad7a9dee5
django_pesapal/views.py
django_pesapal/views.py
# Create your views here. from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs)
# Create your views here. from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
Add support for further processing of the payment while maintaining compatibility
Add support for further processing of the payment while maintaining compatibility
Python
bsd-3-clause
odero/django-pesapal,odero/django-pesapal
# Create your views here. from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) Add support for further processing of the payment while maintaining compatibility
# Create your views here. from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
<commit_before># Create your views here. from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) <commit_msg>Add support for further processing of the payment while maintaining compatibility<commit_after>
# Create your views here. from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
# Create your views here. from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) Add support for further processing of the payment while maintaining compatibility# Create your views here. from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
<commit_before># Create your views here. from django.core.urlresolvers import reverse_lazy from django.contrib.auth.decorators import login_required from django.views.generic.base import RedirectView from django.db.models.loading import get_model from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') merchant_reference = request.GET.get('pesapal_merchant_reference', '') if transaction_id and merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) <commit_msg>Add support for further processing of the payment while maintaining compatibility<commit_after># Create your views here. from django.core.urlresolvers import reverse_lazy, reverse from django.views.generic.base import RedirectView from django.core.urlresolvers import NoReverseMatch from .models import Transaction import conf class TransactionCompletedView(RedirectView): permanent = False url = None def get(self, request, *args, **kwargs): ''' After Pesapal processes the transaction this will save the transaction and then redirect to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`. For further processing just create a `post_save` signal on the `Transaction` model. ''' self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '') self.merchant_reference = request.GET.get('pesapal_merchant_reference', '') if self.transaction_id and self.merchant_reference: transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference, pesapal_transaction=self.transaction_id) return super(TransactionCompletedView, self).get(request, *args, **kwargs) def get_redirect_url(self, *args, **kwargs): try: url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL) except NoReverseMatch: url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL, kwargs={'merchant_reference': self.merchant_reference}) return url
7c1345c09cffaef9710bed24f6ae4ed791c3ade0
setup_cuda_deps.py
setup_cuda_deps.py
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'chainer', 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
Add chainer to required packages
Add chainer to required packages
Python
mit
yanweifu/chainer,ytoyama/yans_chainer_hackathon,AlpacaDB/chainer,jfsantos/chainer,jnishi/chainer,cupy/cupy,woodshop/chainer,okuta/chainer,woodshop/complex-chainer,niboshi/chainer,ysekky/chainer,cupy/cupy,hvy/chainer,ktnyt/chainer,hvy/chainer,wkentaro/chainer,benob/chainer,rezoo/chainer,kiyukuta/chainer,1986ks/chainer,Kaisuke5/chainer,chainer/chainer,minhpqn/chainer,truongdq/chainer,jnishi/chainer,tigerneil/chainer,pfnet/chainer,ikasumi/chainer,benob/chainer,wkentaro/chainer,anaruse/chainer,hvy/chainer,wkentaro/chainer,t-abe/chainer,cupy/cupy,okuta/chainer,delta2323/chainer,okuta/chainer,tkerola/chainer,keisuke-umezawa/chainer,keisuke-umezawa/chainer,muupan/chainer,laysakura/chainer,sinhrks/chainer,ktnyt/chainer,hvy/chainer,chainer/chainer,ktnyt/chainer,kashif/chainer,kuwa32/chainer,umitanuki/chainer,wkentaro/chainer,keisuke-umezawa/chainer,kikusu/chainer,niboshi/chainer,keisuke-umezawa/chainer,t-abe/chainer,kikusu/chainer,sinhrks/chainer,jnishi/chainer,muupan/chainer,niboshi/chainer,tscohen/chainer,elviswf/chainer,AlpacaDB/chainer,truongdq/chainer,cupy/cupy,chainer/chainer,masia02/chainer,wavelets/chainer,hidenori-t/chainer,ronekko/chainer,ktnyt/chainer,okuta/chainer,cemoody/chainer,chainer/chainer,bayerj/chainer,tereka114/chainer,jnishi/chainer,aonotas/chainer,sou81821/chainer,niboshi/chainer
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], ) Add chainer to required packages
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'chainer', 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
<commit_before>#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], ) <commit_msg>Add chainer to required packages<commit_after>
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'chainer', 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], ) Add chainer to required packages#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'chainer', 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
<commit_before>#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], ) <commit_msg>Add chainer to required packages<commit_after>#!/usr/bin/env python from setuptools import setup setup( name='chainer-cuda-deps', version='1.0.0', description='Install dependent packages to use Chainer on CUDA', author='Seiya Tokui', author_email='tokui@preferred.jp', url='http://chainer.org/', packages=[], install_requires=[ 'chainer', 'pycuda>=2014.1', 'scikits.cuda>=0.5.0b1,!=0.042', 'Mako', 'six>=1.9.0', ], )
b7106307baf97ba32cb29fe2a4bb9ed925c194ca
custom/onse/management/commands/update_onse_facility_cases.py
custom/onse/management/commands/update_onse_facility_cases.py
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply( print_notifications=True)
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply(kwargs={ 'print_notifications': True})
Fix passing keyword arg to task
Fix passing keyword arg to task
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply( print_notifications=True) Fix passing keyword arg to task
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply(kwargs={ 'print_notifications': True})
<commit_before>from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply( print_notifications=True) <commit_msg>Fix passing keyword arg to task<commit_after>
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply(kwargs={ 'print_notifications': True})
from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply( print_notifications=True) Fix passing keyword arg to taskfrom django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply(kwargs={ 'print_notifications': True})
<commit_before>from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply( print_notifications=True) <commit_msg>Fix passing keyword arg to task<commit_after>from django.core.management import BaseCommand from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements class Command(BaseCommand): help = ('Update facility_supervision cases with indicators collected ' 'in DHIS2 over the last quarter.') def handle(self, *args, **options): update_facility_cases_from_dhis2_data_elements.apply(kwargs={ 'print_notifications': True})
e5eff6f7f92b2946ca17e59c70b81df6f2e7a12d
opps/core/models/publisher.py
opps/core/models/publisher.py
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True) date_update = models.DateTimeField(_(u"Date update"), auto_now=True) date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() kero = models.Manager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now() def save(self, *args, **kwargs): self.date_update = datetime.now() super(Publisher, self).save(*args, **kwargs)
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now()
Remove date field (insert and update)
Remove date field (insert and update)
Python
mit
jeanmask/opps,jeanmask/opps,jeanmask/opps,YACOWS/opps,opps/opps,williamroot/opps,williamroot/opps,YACOWS/opps,opps/opps,YACOWS/opps,jeanmask/opps,opps/opps,williamroot/opps,opps/opps,YACOWS/opps,williamroot/opps
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True) date_update = models.DateTimeField(_(u"Date update"), auto_now=True) date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() kero = models.Manager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now() def save(self, *args, **kwargs): self.date_update = datetime.now() super(Publisher, self).save(*args, **kwargs) Remove date field (insert and update)
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now()
<commit_before>#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True) date_update = models.DateTimeField(_(u"Date update"), auto_now=True) date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() kero = models.Manager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now() def save(self, *args, **kwargs): self.date_update = datetime.now() super(Publisher, self).save(*args, **kwargs) <commit_msg>Remove date field (insert and update)<commit_after>
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now()
#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True) date_update = models.DateTimeField(_(u"Date update"), auto_now=True) date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() kero = models.Manager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now() def save(self, *args, **kwargs): self.date_update = datetime.now() super(Publisher, self).save(*args, **kwargs) Remove date field (insert and update)#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now()
<commit_before>#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True) date_update = models.DateTimeField(_(u"Date update"), auto_now=True) date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() kero = models.Manager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now() def save(self, *args, **kwargs): self.date_update = datetime.now() super(Publisher, self).save(*args, **kwargs) <commit_msg>Remove date field (insert and update)<commit_after>#!/usr/bin/env python from django.db import models from django.utils.translation import ugettext_lazy as _ from datetime import datetime class PublisherMnager(models.Manager): def all_published(self): return super(PublisherMnager, self).get_query_set().filter( date_available__lte=datetime.now(), published=True) class Publisher(models.Model): date_available = models.DateTimeField(_(u"Date available"), default=datetime.now, null=True) published = models.BooleanField(_(u"Published"), default=False) objects = PublisherMnager() class Meta: abstract = True def is_published(self): return self.published and \ self.date_available.replace(tzinfo=None) <= datetime.now()
1dd58c6717fb8c3c23bce8cecf205c04cc03a134
comrade/views/simple.py
comrade/views/simple.py
from django.http import HttpResponse, HttpResponseServerError from django.template import Context, loader from django.conf import settings import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError( t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
from django.http import HttpResponse, HttpResponseServerError from django.template import RequestContext, loader from django.conf import settings from maintenancemode.http import HttpResponseTemporaryUnavailable import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError(t.render(RequestContext(request))) def maintenance_mode(request, template_name='503.html'): t = loader.get_template(template_name) return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
Use requestcontext in error views.
Use requestcontext in error views.
Python
mit
bueda/django-comrade
from django.http import HttpResponse, HttpResponseServerError from django.template import Context, loader from django.conf import settings import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError( t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL }))) Use requestcontext in error views.
from django.http import HttpResponse, HttpResponseServerError from django.template import RequestContext, loader from django.conf import settings from maintenancemode.http import HttpResponseTemporaryUnavailable import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError(t.render(RequestContext(request))) def maintenance_mode(request, template_name='503.html'): t = loader.get_template(template_name) return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
<commit_before>from django.http import HttpResponse, HttpResponseServerError from django.template import Context, loader from django.conf import settings import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError( t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL }))) <commit_msg>Use requestcontext in error views.<commit_after>
from django.http import HttpResponse, HttpResponseServerError from django.template import RequestContext, loader from django.conf import settings from maintenancemode.http import HttpResponseTemporaryUnavailable import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError(t.render(RequestContext(request))) def maintenance_mode(request, template_name='503.html'): t = loader.get_template(template_name) return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
from django.http import HttpResponse, HttpResponseServerError from django.template import Context, loader from django.conf import settings import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError( t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL }))) Use requestcontext in error views.from django.http import HttpResponse, HttpResponseServerError from django.template import RequestContext, loader from django.conf import settings from maintenancemode.http import HttpResponseTemporaryUnavailable import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError(t.render(RequestContext(request))) def maintenance_mode(request, template_name='503.html'): t = loader.get_template(template_name) return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
<commit_before>from django.http import HttpResponse, HttpResponseServerError from django.template import Context, loader from django.conf import settings import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError( t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL }))) <commit_msg>Use requestcontext in error views.<commit_after>from django.http import HttpResponse, HttpResponseServerError from django.template import RequestContext, loader from django.conf import settings from maintenancemode.http import HttpResponseTemporaryUnavailable import logging logger = logging.getLogger('comrade.views.simple') def status(request): logger.info("Responding to status check") return HttpResponse() def server_error(request, template_name='500.html'): t = loader.get_template(template_name) return HttpResponseServerError(t.render(RequestContext(request))) def maintenance_mode(request, template_name='503.html'): t = loader.get_template(template_name) return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
0ba671698bf4e268ae3f17e11078a5eb669a174c
indico/modules/events/roles/__init__.py
indico/modules/events/roles/__init__.py
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): roles_section = 'organization' if event.type == 'conference' else 'advanced' return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
Move roles menu item into a submenu
Move roles menu item into a submenu - 'organization' for conferences - 'advanced' for other event types
Python
mit
mic4ael/indico,indico/indico,mic4ael/indico,pferreir/indico,DirkHoffmann/indico,OmeGak/indico,pferreir/indico,OmeGak/indico,DirkHoffmann/indico,indico/indico,mvidalgarcia/indico,ThiefMaster/indico,indico/indico,OmeGak/indico,ThiefMaster/indico,DirkHoffmann/indico,DirkHoffmann/indico,mvidalgarcia/indico,ThiefMaster/indico,pferreir/indico,mvidalgarcia/indico,mvidalgarcia/indico,ThiefMaster/indico,pferreir/indico,mic4ael/indico,indico/indico,OmeGak/indico,mic4ael/indico
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal') Move roles menu item into a submenu - 'organization' for conferences - 'advanced' for other event types
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): roles_section = 'organization' if event.type == 'conference' else 'advanced' return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
<commit_before># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal') <commit_msg>Move roles menu item into a submenu - 'organization' for conferences - 'advanced' for other event types<commit_after>
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): roles_section = 'organization' if event.type == 'conference' else 'advanced' return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal') Move roles menu item into a submenu - 'organization' for conferences - 'advanced' for other event types# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): roles_section = 'organization' if event.type == 'conference' else 'advanced' return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
<commit_before># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal') <commit_msg>Move roles menu item into a submenu - 'organization' for conferences - 'advanced' for other event types<commit_after># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from flask import session from indico.core import signals from indico.core.logger import Logger from indico.util.i18n import _ from indico.web.flask.util import url_for from indico.web.menu import SideMenuItem logger = Logger.get('events.roles') @signals.menu.items.connect_via('event-management-sidemenu') def _sidemenu_items(sender, event, **kwargs): if event.can_manage(session.user): roles_section = 'organization' if event.type == 'conference' else 'advanced' return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
8ab4dd1012bedad286e19543f75e5a86d72a0402
biothings/web/api/es/handlers/biothing_handler.py
biothings/web/api/es/handlers/biothing_handler.py
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Inject bid(s) in esqb_kwargs. Set GA tracking object. ''' if self.request.method == 'GET': options.esqb['bid'] = self.path_args[0] # TODO elif self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Set GA tracking object. ''' if self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_query_hook(self, options, query): ''' Request _version field. ''' options.es.version = True return super().pre_query_hook(options, query) def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() # TODO res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
Set _version requirement for annotation
Set _version requirement for annotation
Python
apache-2.0
biothings/biothings.api,biothings/biothings.api
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Inject bid(s) in esqb_kwargs. Set GA tracking object. ''' if self.request.method == 'GET': options.esqb['bid'] = self.path_args[0] # TODO elif self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res Set _version requirement for annotation
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Set GA tracking object. ''' if self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_query_hook(self, options, query): ''' Request _version field. ''' options.es.version = True return super().pre_query_hook(options, query) def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() # TODO res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
<commit_before> from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Inject bid(s) in esqb_kwargs. Set GA tracking object. ''' if self.request.method == 'GET': options.esqb['bid'] = self.path_args[0] # TODO elif self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res <commit_msg>Set _version requirement for annotation<commit_after>
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Set GA tracking object. ''' if self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_query_hook(self, options, query): ''' Request _version field. ''' options.es.version = True return super().pre_query_hook(options, query) def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() # TODO res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Inject bid(s) in esqb_kwargs. Set GA tracking object. ''' if self.request.method == 'GET': options.esqb['bid'] = self.path_args[0] # TODO elif self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res Set _version requirement for annotation from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Set GA tracking object. ''' if self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_query_hook(self, options, query): ''' Request _version field. ''' options.es.version = True return super().pre_query_hook(options, query) def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() # TODO res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
<commit_before> from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Inject bid(s) in esqb_kwargs. Set GA tracking object. ''' if self.request.method == 'GET': options.esqb['bid'] = self.path_args[0] # TODO elif self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res <commit_msg>Set _version requirement for annotation<commit_after> from tornado.web import Finish from biothings.web.api.es.handlers.base_handler import ESRequestHandler class BiothingHandler(ESRequestHandler): ''' Handle requests to the annotation lookup endpoint. ''' name = 'annotation' def pre_query_builder_hook(self, options): ''' Set GA tracking object. ''' if self.request.method == 'POST': self.ga_event_object({'qsize': len(options.esqb.ids)}) return options def pre_query_hook(self, options, query): ''' Request _version field. ''' options.es.version = True return super().pre_query_hook(options, query) def pre_finish_hook(self, options, res): ''' Return single result for GET. Also does not return empty results. ''' if isinstance(res, dict): if not res.get('hits'): self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid)) raise Finish() # TODO res = res['hits'][0] res.pop('_score') elif isinstance(res, list): for hit in res: hit.pop('_score', None) return res
369e70cc5d5e1c880bd61f055ffbe43d5aeab509
nodeconductor/structure/tests/unittests/test_models.py
nodeconductor/structure/tests/unittests/test_models.py
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
Break test flow into logical subsets
Break test flow into logical subsets [WAL-615]
Python
mit
opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) Break test flow into logical subsets [WAL-615]
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
<commit_before>from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) <commit_msg>Break test flow into logical subsets [WAL-615]<commit_after>
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) Break test flow into logical subsets [WAL-615]from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
<commit_before>from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) <commit_msg>Break test flow into logical subsets [WAL-615]<commit_after>from django.test import TestCase from nodeconductor.structure.tests import factories class ServiceProjectLinkTest(TestCase): def setUp(self): self.link = factories.TestServiceProjectLinkFactory() def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self): certification = factories.ServiceCertificationFactory() self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant) self.link.project.certifications.add(certification) self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant) def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self): certifications = factories.ServiceCertificationFactory.create_batch(2) self.link.project.certifications.add(*certifications) certifications.append(factories.ServiceCertificationFactory()) self.link.service.settings.certifications.add(*certifications) self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
5a785f725d68733561a7e5e82c57655e25439ec8
indra/tests/test_grounding_resources.py
indra/tests/test_grounding_resources.py
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
Remove exceptional namespaces from test
Remove exceptional namespaces from test
Python
bsd-2-clause
johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/indra,bgyori/indra,sorgerlab/indra,bgyori/indra,bgyori/indra,johnbachman/indra,johnbachman/indra,sorgerlab/belpy,sorgerlab/belpy
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled Remove exceptional namespaces from test
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
<commit_before>import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled <commit_msg>Remove exceptional namespaces from test<commit_after>
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled Remove exceptional namespaces from testimport os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
<commit_before>import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map # Namespaces that are not currently handled but still appear in statements exceptions = ['CLO'] def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_exceptional_unhandled(): """Test that exceptional namespaces actually aren't handled. This will catch if we make an update that makes an exceptional namespace become a handled namespace. That way we can update the tests. """ actually_handled = [] for ns in exceptions: if validate_ns(ns): actually_handled.append(ns) assert not actually_handled, actually_handled <commit_msg>Remove exceptional namespaces from test<commit_after>import os import csv from indra.statements.validate import validate_db_refs, validate_ns from indra.preassembler.grounding_mapper import default_grounding_map from indra.preassembler.grounding_mapper import default_misgrounding_map def test_misgrounding_map_entries(): bad_entries = [] for text, db_refs in default_misgrounding_map.items(): if not validate_db_refs(db_refs): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries def test_grounding_map_entries(): bad_entries = [] for text, db_refs in default_grounding_map.items(): if (not validate_db_refs(db_refs) and not (set(exceptions) & db_refs.keys())): bad_entries.append([text, db_refs]) assert not bad_entries, bad_entries
1e42bc1ef04ff3f52ce3f5db75d781be7d450a25
etl_framework/etl_class.py
etl_framework/etl_class.py
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.config return getattr(config, key)
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif "config" in self.__dict__ and hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.__dict__["config"] return getattr(config, key)
Make EtlClass attribute access more robust
Make EtlClass attribute access more robust
Python
mit
pantheon-systems/etl-framework
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.config return getattr(config, key) Make EtlClass attribute access more robust
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif "config" in self.__dict__ and hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.__dict__["config"] return getattr(config, key)
<commit_before>"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.config return getattr(config, key) <commit_msg>Make EtlClass attribute access more robust<commit_after>
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif "config" in self.__dict__ and hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.__dict__["config"] return getattr(config, key)
"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.config return getattr(config, key) Make EtlClass attribute access more robust"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif "config" in self.__dict__ and hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.__dict__["config"] return getattr(config, key)
<commit_before>"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.config return getattr(config, key) <commit_msg>Make EtlClass attribute access more robust<commit_after>"""Base EtlClass that all EtlClasses should inherit""" class EtlClass(object): def __init__(self, config): self.config = config def __setattr__(self, key, value): """Set attribute on config if not in EtlClass object""" if key == "config": self.__dict__[key] = value elif "config" in self.__dict__ and hasattr(self.config, key): setattr(self.config, key, value) else: self.__dict__[key] = value def __getattr__(self, key): """Get attribute on config if not in EtlClass object""" # Get attribute if Config doesnt exist # we don't need a special call to super here because getattr is only # called when an attribute is NOT found in the instance's dictionary config = self.__dict__["config"] return getattr(config, key)
1f83113e748963cda9688d88a5d36dd7f9a54c1f
tests/app/test_cloudfoundry_config.py
tests/app/test_cloudfoundry_config.py
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def cloudfoundry_environ(os_environ, cloudfoundry_config): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ): extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def vcap_application(os_environ): os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
Move setting VCAP_SERVICES out of fixture
Move setting VCAP_SERVICES out of fixture This was inconsistent with the source data for the fixture being overidden in some of the tests. We only need to set it in the env once, so it makes sense to just put the code there.
Python
mit
alphagov/notifications-api,alphagov/notifications-api
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def cloudfoundry_environ(os_environ, cloudfoundry_config): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ): extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ Move setting VCAP_SERVICES out of fixture This was inconsistent with the source data for the fixture being overidden in some of the tests. We only need to set it in the env once, so it makes sense to just put the code there.
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def vcap_application(os_environ): os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
<commit_before>import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def cloudfoundry_environ(os_environ, cloudfoundry_config): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ): extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ <commit_msg>Move setting VCAP_SERVICES out of fixture This was inconsistent with the source data for the fixture being overidden in some of the tests. We only need to set it in the env once, so it makes sense to just put the code there.<commit_after>
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def vcap_application(os_environ): os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def cloudfoundry_environ(os_environ, cloudfoundry_config): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ): extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ Move setting VCAP_SERVICES out of fixture This was inconsistent with the source data for the fixture being overidden in some of the tests. We only need to set it in the env once, so it makes sense to just put the code there.import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def vcap_application(os_environ): os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
<commit_before>import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def cloudfoundry_environ(os_environ, cloudfoundry_config): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ): extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ <commit_msg>Move setting VCAP_SERVICES out of fixture This was inconsistent with the source data for the fixture being overidden in some of the tests. We only need to set it in the env once, so it makes sense to just put the code there.<commit_after>import json import os import pytest from app.cloudfoundry_config import ( extract_cloudfoundry_config, set_config_env_vars, ) @pytest.fixture def cloudfoundry_config(): return { 'postgres': [{ 'credentials': { 'uri': 'postgres uri' } }], 'user-provided': [] } @pytest.fixture def vcap_application(os_environ): os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}' def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application): os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config) extract_cloudfoundry_config() assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri' assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌' assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log' def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application): cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}} cloudfoundry_config['user-provided'].append({ 'name': 'bar', 'credentials': {'bar': 'bar'} }) set_config_env_vars(cloudfoundry_config) assert 'foo' not in os.environ assert 'bar' not in os.environ
7cbe2351c2ad93def98005597a24e21d878ea492
flask_velox/mixins/http.py
flask_velox/mixins/http.py
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self, *args, **kwargs): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
Allow RedirectMixin to work within flask-admin
Allow RedirectMixin to work within flask-admin
Python
mit
thisissoon/Flask-Velox,thisissoon/Flask-Velox,jstacoder/Flask-Velox,jstacoder/Flask-Velox
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302)) Allow RedirectMixin to work within flask-admin
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self, *args, **kwargs): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
<commit_before># -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302)) <commit_msg>Allow RedirectMixin to work within flask-admin<commit_after>
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self, *args, **kwargs): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302)) Allow RedirectMixin to work within flask-admin# -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self, *args, **kwargs): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
<commit_before># -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302)) <commit_msg>Allow RedirectMixin to work within flask-admin<commit_after># -*- coding: utf-8 -*- """ Module provides mixins for issuing HTTP Status codes using the Flask ``View``. """ from flask import url_for from flask.views import View from werkzeug.utils import redirect class RedirectMixin(View): """ Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used however this can be overridden using the ``code`` attribute. Example ------- .. code-block:: python :linenos: from flask.ext.velox.mixins.http import RedirectMixin class MyView(RedirectMixin): rule = 'some.url.rule' code = 301 Attributes ---------- rule : str Flask URL Rule passed into ``url_for`` code : int, optional Status code to raise, defaults to ``302`` """ code = 302 def pre_dispatch(self, *args, **kwargs): """ If you wish to run an arbitrary piece of code before the redirect is dispatched you can override this method which is called before dispatch. """ pass def get_url(self): """ Return a generated url from ``rule`` attribute. Returns ------- str Generated url """ try: rule = self.rule except AttributeError: raise NotImplementedError('``rule`` attr must be defined.') return url_for(rule) def dispatch_request(self, *args, **kwargs): """ Dispatch the request, returning the redirect.func_closure Returns ------- werkzeug.wrappers.Response Redirect response """ self.pre_dispatch() return redirect(self.get_url(), code=getattr(self, 'code', 302))
c058ea00594efcdb39bb106ebe79dde39354e159
downloader.py
downloader.py
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "0", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "9", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
Use low quality audio (sounds good enough)
Use low quality audio (sounds good enough)
Python
mit
xrisk/webdl,xrisk/webdl,xrisk/webdl
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "0", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception Use low quality audio (sounds good enough)
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "9", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
<commit_before>import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "0", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception <commit_msg>Use low quality audio (sounds good enough)<commit_after>
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "9", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "0", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception Use low quality audio (sounds good enough)import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "9", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
<commit_before>import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "0", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception <commit_msg>Use low quality audio (sounds good enough)<commit_after>import sha import os def sha_hash(content): return sha.new(content).hexdigest() def download_audio(url): from subprocess import call retcode = call(["youtube-dl", "-x", "--id", "--audio-quality", "9", "--audio-format", "mp3", "--exec", "mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception def download_video(url): from subprocess import call retcode = call(["youtube-dl", "--format", "mp4", "--exec", "mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)), url]) if retcode == 0: return sha_hash(url) else: raise Exception
bdda5e565981ac26a7e5e1ab8d1486eb91b09e4c
views/base.py
views/base.py
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path): page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path=None): if path is None: path = request.path page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
Use request.path if no path was passed to the default view
Use request.path if no path was passed to the default view
Python
bsd-3-clause
mjl/feincms,nickburlett/feincms,nickburlett/feincms,pjdelport/feincms,matthiask/django-content-editor,michaelkuty/feincms,matthiask/feincms2-content,mjl/feincms,matthiask/django-content-editor,matthiask/django-content-editor,mjl/feincms,hgrimelid/feincms,hgrimelid/feincms,joshuajonah/feincms,nickburlett/feincms,matthiask/django-content-editor,joshuajonah/feincms,joshuajonah/feincms,pjdelport/feincms,michaelkuty/feincms,matthiask/feincms2-content,feincms/feincms,pjdelport/feincms,feincms/feincms,michaelkuty/feincms,nickburlett/feincms,joshuajonah/feincms,feincms/feincms,matthiask/feincms2-content,michaelkuty/feincms,hgrimelid/feincms
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path): page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request)) Use request.path if no path was passed to the default view
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path=None): if path is None: path = request.path page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
<commit_before>from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path): page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request)) <commit_msg>Use request.path if no path was passed to the default view<commit_after>
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path=None): if path is None: path = request.path page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path): page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request)) Use request.path if no path was passed to the default viewfrom django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path=None): if path is None: path = request.path page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
<commit_before>from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path): page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request)) <commit_msg>Use request.path if no path was passed to the default view<commit_after>from django.http import HttpResponseRedirect from django.shortcuts import render_to_response from django.template import RequestContext from feincms.models import Page def handler(request, path=None): if path is None: path = request.path page = Page.objects.page_for_path_or_404(path) if page.override_url: return HttpResponseRedirect(page.override_url) return render_to_response(page.template.path, { 'feincms_page': page, }, context_instance=RequestContext(request))
31bfe8fb498ea2e528da6463c9045b397992e028
python/caffe/test/test_draw.py
python/caffe/test/test_draw.py
import os import unittest from google import protobuf import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: protobuf.text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR')
import os import unittest from google.protobuf import text_format import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') if __name__ == "__main__": unittest.main()
Add main() for draw_net unittest, fix import errors
Add main() for draw_net unittest, fix import errors
Python
apache-2.0
gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina
import os import unittest from google import protobuf import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: protobuf.text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') Add main() for draw_net unittest, fix import errors
import os import unittest from google.protobuf import text_format import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') if __name__ == "__main__": unittest.main()
<commit_before>import os import unittest from google import protobuf import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: protobuf.text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') <commit_msg>Add main() for draw_net unittest, fix import errors<commit_after>
import os import unittest from google.protobuf import text_format import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') if __name__ == "__main__": unittest.main()
import os import unittest from google import protobuf import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: protobuf.text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') Add main() for draw_net unittest, fix import errorsimport os import unittest from google.protobuf import text_format import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') if __name__ == "__main__": unittest.main()
<commit_before>import os import unittest from google import protobuf import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: protobuf.text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') <commit_msg>Add main() for draw_net unittest, fix import errors<commit_after>import os import unittest from google.protobuf import text_format import caffe.draw from caffe.proto import caffe_pb2 def getFilenames(): """Yields files in the source tree which are Net prototxts.""" result = [] root_dir = os.path.abspath(os.path.join( os.path.dirname(__file__), '..', '..', '..')) assert os.path.exists(root_dir) for dirname in ('models', 'examples'): dirname = os.path.join(root_dir, dirname) assert os.path.exists(dirname) for cwd, _, filenames in os.walk(dirname): for filename in filenames: filename = os.path.join(cwd, filename) if filename.endswith('.prototxt') and 'solver' not in filename: yield os.path.join(dirname, filename) class TestDraw(unittest.TestCase): def test_draw_net(self): for filename in getFilenames(): net = caffe_pb2.NetParameter() with open(filename) as infile: text_format.Merge(infile.read(), net) caffe.draw.draw_net(net, 'LR') if __name__ == "__main__": unittest.main()
b64bac9def90d53e76a3427f04f8664d513ae7ed
wallace/db.py
wallace/db.py
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
Fix assignment of database url
Fix assignment of database url
Python
mit
berkeley-cocosci/Wallace,Dallinger/Dallinger,berkeley-cocosci/Wallace,jcpeterson/Dallinger,suchow/Wallace,berkeley-cocosci/Wallace,Dallinger/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,suchow/Wallace,suchow/Wallace,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session Fix assignment of database url
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
<commit_before>"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session <commit_msg>Fix assignment of database url<commit_after>
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session Fix assignment of database url"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
<commit_before>"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session <commit_msg>Fix assignment of database url<commit_after>"""Create a connection to the database.""" from sqlalchemy import create_engine from sqlalchemy.orm import sessionmaker, scoped_session from sqlalchemy.ext.declarative import declarative_base import os db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace") engine = create_engine(db_url) Session = scoped_session(sessionmaker(autoflush=True, bind=engine)) Base = declarative_base() Base.query = Session.query_property() def init_db(drop_all=False): """Initialize the database, optionally dropping existing tables.""" if drop_all: Base.metadata.drop_all(bind=engine) Base.metadata.create_all(bind=engine) return Session
e452663965f63ffb1d1e3ac04ae10dc26b3e2ecc
fabric/api.py
fabric/api.py
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env from fabric.utils import abort, warn
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env, output from fabric.utils import abort, warn
Allow access to state.output via API.
Allow access to state.output via API.
Python
bsd-2-clause
bitmonk/fabric,fernandezcuesta/fabric,opavader/fabric,xLegoz/fabric,getsentry/fabric,bitprophet/fabric,jaraco/fabric,MjAbuz/fabric,SamuelMarks/fabric,pashinin/fabric,tolbkni/fabric,pgroudas/fabric,bspink/fabric,likesxuqiang/fabric,hrubi/fabric,TarasRudnyk/fabric,sdelements/fabric,tekapo/fabric,simon-engledew/fabric,ericholscher/fabric,askulkarni2/fabric,ploxiln/fabric,kxxoling/fabric,haridsv/fabric,amaniak/fabric,kmonsoor/fabric,rodrigc/fabric,qinrong/fabric,cmattoon/fabric,cgvarela/fabric,raimon49/fabric,mathiasertl/fabric,akaariai/fabric,elijah513/fabric,rane-hs/fabric-py3,rbramwell/fabric,felix-d/fabric,itoed/fabric,StackStorm/fabric
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env from fabric.utils import abort, warn Allow access to state.output via API.
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env, output from fabric.utils import abort, warn
<commit_before>""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env from fabric.utils import abort, warn <commit_msg>Allow access to state.output via API.<commit_after>
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env, output from fabric.utils import abort, warn
""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env from fabric.utils import abort, warn Allow access to state.output via API.""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env, output from fabric.utils import abort, warn
<commit_before>""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env from fabric.utils import abort, warn <commit_msg>Allow access to state.output via API.<commit_after>""" Non-init module for doing convenient * imports from. Necessary because if we did this in __init__, one would be unable to import anything else inside the package -- like, say, the version number used in setup.py -- without triggering loads of most of the code. Which doesn't work so well when you're using setup.py to install e.g. paramiko! """ from fabric.context_managers import cd, hide, settings, show from fabric.decorators import hosts, roles, runs_once from fabric.operations import require, prompt, put, get, run, sudo, local from fabric.state import env, output from fabric.utils import abort, warn
d1e1f63062eff158b9bce8b9c3cbcaef1abed8ba
flask_gzip.py
flask_gzip.py
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers: return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if response.direct_passthrough: return response if (response.status_code not in xrange(200, 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers): return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
Fix a runtime error when direct_passthrough is used.
Fix a runtime error when direct_passthrough is used.
Python
mit
wichitacode/flask-compress,wichitacode/flask-compress,libwilliam/flask-compress,saymedia/flask-compress,saymedia/flask-compress,libwilliam/flask-compress,libwilliam/flask-compress
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers: return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response Fix a runtime error when direct_passthrough is used.
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if response.direct_passthrough: return response if (response.status_code not in xrange(200, 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers): return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
<commit_before>import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers: return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response <commit_msg>Fix a runtime error when direct_passthrough is used.<commit_after>
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if response.direct_passthrough: return response if (response.status_code not in xrange(200, 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers): return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers: return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response Fix a runtime error when direct_passthrough is used.import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if response.direct_passthrough: return response if (response.status_code not in xrange(200, 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers): return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
<commit_before>import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers: return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response <commit_msg>Fix a runtime error when direct_passthrough is used.<commit_after>import gzip import StringIO from flask import request class Gzip(object): def __init__(self, app, compress_level=6, minimum_size=500): self.app = app self.compress_level = compress_level self.minimum_size = minimum_size self.app.after_request(self.after_request) def after_request(self, response): accept_encoding = request.headers.get('Accept-Encoding', '') if 'gzip' not in accept_encoding.lower(): return response if response.direct_passthrough: return response if (response.status_code not in xrange(200, 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers): return response gzip_buffer = StringIO.StringIO() gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer) gzip_file.write(response.data) gzip_file.close() response.data = gzip_buffer.getvalue() response.headers['Content-Encoding'] = 'gzip' response.headers['Content-Length'] = len(response.data) return response
23ad531d932b6c042c3bd0161b74a6088d02524f
myfedora/lib/app_globals.py
myfedora/lib/app_globals.py
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets from myfedora.streams import DataStreamer self.datastreamer = DataStreamer()
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets #from myfedora.streams import DataStreamer #self.datastreamer = DataStreamer() FEED_CACHE = "/tmp/moksha-feeds" from shove import Shove from feedcache.cache import Cache # is this not multi-process safe? or even thread safe? self.feed_storage = Shove('file://' + FEED_CACHE) self.feed_cache = Cache(self.feed_storage)
Add a feed_storage and feed_cache to our Globals object.
Add a feed_storage and feed_cache to our Globals object.
Python
agpl-3.0
fedora-infra/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages,Fale/fedora-packages,fedora-infra/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets from myfedora.streams import DataStreamer self.datastreamer = DataStreamer() Add a feed_storage and feed_cache to our Globals object.
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets #from myfedora.streams import DataStreamer #self.datastreamer = DataStreamer() FEED_CACHE = "/tmp/moksha-feeds" from shove import Shove from feedcache.cache import Cache # is this not multi-process safe? or even thread safe? self.feed_storage = Shove('file://' + FEED_CACHE) self.feed_cache = Cache(self.feed_storage)
<commit_before>"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets from myfedora.streams import DataStreamer self.datastreamer = DataStreamer() <commit_msg>Add a feed_storage and feed_cache to our Globals object.<commit_after>
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets #from myfedora.streams import DataStreamer #self.datastreamer = DataStreamer() FEED_CACHE = "/tmp/moksha-feeds" from shove import Shove from feedcache.cache import Cache # is this not multi-process safe? or even thread safe? self.feed_storage = Shove('file://' + FEED_CACHE) self.feed_cache = Cache(self.feed_storage)
"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets from myfedora.streams import DataStreamer self.datastreamer = DataStreamer() Add a feed_storage and feed_cache to our Globals object."""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets #from myfedora.streams import DataStreamer #self.datastreamer = DataStreamer() FEED_CACHE = "/tmp/moksha-feeds" from shove import Shove from feedcache.cache import Cache # is this not multi-process safe? or even thread safe? self.feed_storage = Shove('file://' + FEED_CACHE) self.feed_cache = Cache(self.feed_storage)
<commit_before>"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets from myfedora.streams import DataStreamer self.datastreamer = DataStreamer() <commit_msg>Add a feed_storage and feed_cache to our Globals object.<commit_after>"""The application's Globals object""" from app_factory import AppFactoryDict class Globals(object): """Globals acts as a container for objects available throughout the life of the application """ def __init__(self): """One instance of Globals is created during application initialization and is available during requests via the 'g' variable """ self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}} self.resourceviews = AppFactoryDict() # {name: ResourceView instance} self.apps = AppFactoryDict() # {name: App instance} # Our comet data streamer, responsible for polling the data # streams, and providing data to the widgets #from myfedora.streams import DataStreamer #self.datastreamer = DataStreamer() FEED_CACHE = "/tmp/moksha-feeds" from shove import Shove from feedcache.cache import Cache # is this not multi-process safe? or even thread safe? self.feed_storage = Shove('file://' + FEED_CACHE) self.feed_cache = Cache(self.feed_storage)
e9f3b6f9eb59ef7290498e8ceaf81c2bc66c8f59
ichnaea/gunicorn_config.py
ichnaea/gunicorn_config.py
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). # It should be 60 seconds, but gunicorn halves the configured value, # see https://github.com/benoitc/gunicorn/issues/829 timeout = 120 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). timeout = 60 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
Update gunicorn timeout after gunicorn issue was answered.
Update gunicorn timeout after gunicorn issue was answered.
Python
apache-2.0
mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). # It should be 60 seconds, but gunicorn halves the configured value, # see https://github.com/benoitc/gunicorn/issues/829 timeout = 120 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None) Update gunicorn timeout after gunicorn issue was answered.
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). timeout = 60 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
<commit_before># This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). # It should be 60 seconds, but gunicorn halves the configured value, # see https://github.com/benoitc/gunicorn/issues/829 timeout = 120 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None) <commit_msg>Update gunicorn timeout after gunicorn issue was answered.<commit_after>
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). timeout = 60 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). # It should be 60 seconds, but gunicorn halves the configured value, # see https://github.com/benoitc/gunicorn/issues/829 timeout = 120 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None) Update gunicorn timeout after gunicorn issue was answered.# This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). timeout = 60 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
<commit_before># This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). # It should be 60 seconds, but gunicorn halves the configured value, # see https://github.com/benoitc/gunicorn/issues/829 timeout = 120 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None) <commit_msg>Update gunicorn timeout after gunicorn issue was answered.<commit_after># This file contains gunicorn configuration setttings, as described at # http://docs.gunicorn.org/en/latest/settings.html # The file is loaded via the -c ichnaea.gunicorn_config command line option # Be explicit about the worker class worker_class = "sync" # Set timeout to the same value as the default one from Amazon ELB (60 secs). timeout = 60 # Recycle worker processes after 100k requests to prevent memory leaks # from effecting us max_requests = 100000 # Avoid too much output on the console loglevel = "warning" def post_worker_init(worker): from random import randint # Use 10% jitter, to prevent all workers from restarting at once, # as they get an almost equal number of requests jitter = randint(0, max_requests // 10) worker.max_requests += jitter # Actually initialize the application worker.wsgi(None, None)
1c8f29d78d6409ba58df36d439f1ffd436c9dd10
gaphas/picklers.py
gaphas/picklers.py
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import new def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return new.instancemethod(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import types def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return types.MethodType(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
Fix no module 'new', replaced new.instancemethod with types.MethodType
Fix no module 'new', replaced new.instancemethod with types.MethodType Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
Python
lgpl-2.1
amolenaar/gaphas
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import new def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return new.instancemethod(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai Fix no module 'new', replaced new.instancemethod with types.MethodType Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import types def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return types.MethodType(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
<commit_before>""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import new def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return new.instancemethod(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai <commit_msg>Fix no module 'new', replaced new.instancemethod with types.MethodType Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import types def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return types.MethodType(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import new def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return new.instancemethod(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai Fix no module 'new', replaced new.instancemethod with types.MethodType Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import types def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return types.MethodType(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
<commit_before>""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import new def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return new.instancemethod(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai <commit_msg>Fix no module 'new', replaced new.instancemethod with types.MethodType Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>""" Some extra picklers needed to gracefully dump and load a canvas. """ from future import standard_library standard_library.install_aliases() import copyreg # Allow instancemethod to be pickled: import types def construct_instancemethod(funcname, self, clazz): func = getattr(clazz, funcname) return types.MethodType(func, self, clazz) def reduce_instancemethod(im): return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__) copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod) # Allow cairo.Matrix to be pickled: import cairo def construct_cairo_matrix(*args): return cairo.Matrix(*args) def reduce_cairo_matrix(m): return construct_cairo_matrix, tuple(m) copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix) # vim:sw=4:et:ai
62f6e4a07ef63d53e1ff4230b7a40ad38168d103
python/testData/highlighting/assignmentTargets3K.py
python/testData/highlighting/assignmentTargets3K.py
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="assignment to keyword">True</error> = 2 <error descr="assignment to keyword">False</error> = 1 <error descr="assignment to keyword">None</error> = 0 <error descr="assignment to keyword">__debug__</error> = -1 () = [] [] = []
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="Assignment to keyword">True</error> = 2 <error descr="Assignment to keyword">False</error> = 1 <error descr="Assignment to keyword">None</error> = 0 <error descr="Assignment to keyword">__debug__</error> = -1 () = [] [] = []
Fix test broken by change in message capitalization
Fix test broken by change in message capitalization GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda
Python
apache-2.0
allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="assignment to keyword">True</error> = 2 <error descr="assignment to keyword">False</error> = 1 <error descr="assignment to keyword">None</error> = 0 <error descr="assignment to keyword">__debug__</error> = -1 () = [] [] = [] Fix test broken by change in message capitalization GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="Assignment to keyword">True</error> = 2 <error descr="Assignment to keyword">False</error> = 1 <error descr="Assignment to keyword">None</error> = 0 <error descr="Assignment to keyword">__debug__</error> = -1 () = [] [] = []
<commit_before><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="assignment to keyword">True</error> = 2 <error descr="assignment to keyword">False</error> = 1 <error descr="assignment to keyword">None</error> = 0 <error descr="assignment to keyword">__debug__</error> = -1 () = [] [] = [] <commit_msg>Fix test broken by change in message capitalization GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<commit_after>
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="Assignment to keyword">True</error> = 2 <error descr="Assignment to keyword">False</error> = 1 <error descr="Assignment to keyword">None</error> = 0 <error descr="Assignment to keyword">__debug__</error> = -1 () = [] [] = []
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="assignment to keyword">True</error> = 2 <error descr="assignment to keyword">False</error> = 1 <error descr="assignment to keyword">None</error> = 0 <error descr="assignment to keyword">__debug__</error> = -1 () = [] [] = [] Fix test broken by change in message capitalization GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="Assignment to keyword">True</error> = 2 <error descr="Assignment to keyword">False</error> = 1 <error descr="Assignment to keyword">None</error> = 0 <error descr="Assignment to keyword">__debug__</error> = -1 () = [] [] = []
<commit_before><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="assignment to keyword">True</error> = 2 <error descr="assignment to keyword">False</error> = 1 <error descr="assignment to keyword">None</error> = 0 <error descr="assignment to keyword">__debug__</error> = -1 () = [] [] = [] <commit_msg>Fix test broken by change in message capitalization GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<commit_after><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2] *a, = range(5) for a, *b in [(1, 2, 3), (4, 5, 6, 7)]: print(b) a, b, c = seq[0], seq[1:-1], seq[-1] a, *b, c = seq [a, *b, c] = seq <error descr="Assignment to keyword">True</error> = 2 <error descr="Assignment to keyword">False</error> = 1 <error descr="Assignment to keyword">None</error> = 0 <error descr="Assignment to keyword">__debug__</error> = -1 () = [] [] = []
48c9b0fc46da538633e7597bb919ac15e4accf7c
zeus/db/func.py
zeus/db/func.py
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] padding = tuple((len(self.cols) - len(elems)) * (None,)) return [tuple(self._param_regexp.findall(e)) + padding for e in elems] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] num_cols = len(self.cols) padding = (None,) * num_cols return [ (tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems ] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
Correct padding on array aggregations
fix: Correct padding on array aggregations This was incorrectly building padding based on a single row.
Python
apache-2.0
getsentry/zeus,getsentry/zeus,getsentry/zeus,getsentry/zeus
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] padding = tuple((len(self.cols) - len(elems)) * (None,)) return [tuple(self._param_regexp.findall(e)) + padding for e in elems] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg)) fix: Correct padding on array aggregations This was incorrectly building padding based on a single row.
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] num_cols = len(self.cols) padding = (None,) * num_cols return [ (tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems ] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
<commit_before>import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] padding = tuple((len(self.cols) - len(elems)) * (None,)) return [tuple(self._param_regexp.findall(e)) + padding for e in elems] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg)) <commit_msg>fix: Correct padding on array aggregations This was incorrectly building padding based on a single row.<commit_after>
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] num_cols = len(self.cols) padding = (None,) * num_cols return [ (tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems ] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] padding = tuple((len(self.cols) - len(elems)) * (None,)) return [tuple(self._param_regexp.findall(e)) + padding for e in elems] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg)) fix: Correct padding on array aggregations This was incorrectly building padding based on a single row.import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] num_cols = len(self.cols) padding = (None,) * num_cols return [ (tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems ] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
<commit_before>import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] padding = tuple((len(self.cols) - len(elems)) * (None,)) return [tuple(self._param_regexp.findall(e)) + padding for e in elems] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg)) <commit_msg>fix: Correct padding on array aggregations This was incorrectly building padding based on a single row.<commit_after>import re from sqlalchemy.sql import func from sqlalchemy.types import String, TypeDecorator # https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does class ArrayOfRecord(TypeDecorator): _array_regexp = re.compile(r"^\{(\".+?\")*\}$") _chunk_regexp = re.compile(r'"(.*?)",?') _param_regexp = re.compile(r"[^\(\),]+") impl = String def __init__(self, cols): self.cols = cols super().__init__() def process_result_value(self, value, dialect): # XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to # not return them, and thus our output array does not match the same length as our column # selection array # # For example if the input is: # ARRAY_AGG_RESULT(col1, col2) # And the value of col2 is NULL # The resulting return value from this query will be: # ({col1_value},) elems = self._array_regexp.match(value).group(1) elems = [e for e in self._chunk_regexp.split(elems) if e] num_cols = len(self.cols) padding = (None,) * num_cols return [ (tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems ] def array_agg_row(*arg): return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
83042027fe74ffe200d0bdaa79b0529af54ae6dc
addons/website/__openerp__.py
addons/website/__openerp__.py
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['base_setup', 'web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
Revert "[FIX] website: add missing module dependency `base_setup`"
Revert "[FIX] website: add missing module dependency `base_setup`" This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1. The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.
Python
agpl-3.0
ygol/odoo,dfang/odoo,hip-odoo/odoo,hip-odoo/odoo,ygol/odoo,ygol/odoo,dfang/odoo,dfang/odoo,ygol/odoo,hip-odoo/odoo,hip-odoo/odoo,ygol/odoo,ygol/odoo,ygol/odoo,dfang/odoo,hip-odoo/odoo,hip-odoo/odoo,dfang/odoo,dfang/odoo
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['base_setup', 'web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, } Revert "[FIX] website: add missing module dependency `base_setup`" This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1. The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
<commit_before># -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['base_setup', 'web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, } <commit_msg>Revert "[FIX] website: add missing module dependency `base_setup`" This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1. The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.<commit_after>
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['base_setup', 'web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, } Revert "[FIX] website: add missing module dependency `base_setup`" This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1. The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.# -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
<commit_before># -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['base_setup', 'web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, } <commit_msg>Revert "[FIX] website: add missing module dependency `base_setup`" This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1. The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.<commit_after># -*- encoding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Website Builder', 'category': 'Website', 'sequence': 50, 'summary': 'Build Your Enterprise Website', 'website': 'https://www.odoo.com/page/website-builder', 'version': '1.0', 'description': """ Odoo Website CMS =================== """, 'depends': ['web', 'web_editor', 'web_planner'], 'installable': True, 'data': [ 'data/website_data.xml', 'data/web_planner_data.xml', 'security/ir.model.access.csv', 'security/website_security.xml', 'views/website_templates.xml', 'views/website_navbar_templates.xml', 'views/snippets.xml', 'views/website_views.xml', 'views/res_config_views.xml', 'views/ir_actions_views.xml', 'wizard/base_language_install_views.xml', ], 'demo': [ 'data/website_demo.xml', ], 'qweb': ['static/src/xml/website.backend.xml'], 'application': True, }
c138adaf69f5029209f03cafe72f1082cdb78f30
ppp_nlp_ml_standalone/requesthandler.py
ppp_nlp_ml_standalone/requesthandler.py
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() a, b, c = extract_triplet.extract_from_sentence(sentence) if a == '?': subject = ppp_datamodel.Missing() else: subject = ppp_datamodel.Resource(value=a) if b == '?': predicate = ppp_datamodel.Missing() else: predicate = ppp_datamodel.Resource(value=b) if c == '?': object = ppp_datamodel.Missing() else: object = ppp_datamodel.Resource(value=b) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence, Missing, Resource from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet def missing_or_resource(x): return Missing() if x == '?' else Resource(value=x) class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() triple = extract_triplet.extract_from_sentence(sentence) (subject, predicate, object) = map(missing_or_resource, triple) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
Make RequestHandler's code less redundant.
Make RequestHandler's code less redundant.
Python
mit
ProjetPP/PPP-QuestionParsing-ML-Standalone,ProjetPP/PPP-QuestionParsing-ML-Standalone
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() a, b, c = extract_triplet.extract_from_sentence(sentence) if a == '?': subject = ppp_datamodel.Missing() else: subject = ppp_datamodel.Resource(value=a) if b == '?': predicate = ppp_datamodel.Missing() else: predicate = ppp_datamodel.Resource(value=b) if c == '?': object = ppp_datamodel.Missing() else: object = ppp_datamodel.Resource(value=b) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]Make RequestHandler's code less redundant.
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence, Missing, Resource from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet def missing_or_resource(x): return Missing() if x == '?' else Resource(value=x) class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() triple = extract_triplet.extract_from_sentence(sentence) (subject, predicate, object) = map(missing_or_resource, triple) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
<commit_before>"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() a, b, c = extract_triplet.extract_from_sentence(sentence) if a == '?': subject = ppp_datamodel.Missing() else: subject = ppp_datamodel.Resource(value=a) if b == '?': predicate = ppp_datamodel.Missing() else: predicate = ppp_datamodel.Resource(value=b) if c == '?': object = ppp_datamodel.Missing() else: object = ppp_datamodel.Resource(value=b) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]<commit_msg>Make RequestHandler's code less redundant.<commit_after>
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence, Missing, Resource from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet def missing_or_resource(x): return Missing() if x == '?' else Resource(value=x) class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() triple = extract_triplet.extract_from_sentence(sentence) (subject, predicate, object) = map(missing_or_resource, triple) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() a, b, c = extract_triplet.extract_from_sentence(sentence) if a == '?': subject = ppp_datamodel.Missing() else: subject = ppp_datamodel.Resource(value=a) if b == '?': predicate = ppp_datamodel.Missing() else: predicate = ppp_datamodel.Resource(value=b) if c == '?': object = ppp_datamodel.Missing() else: object = ppp_datamodel.Resource(value=b) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]Make RequestHandler's code less redundant."""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence, Missing, Resource from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet def missing_or_resource(x): return Missing() if x == '?' else Resource(value=x) class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() triple = extract_triplet.extract_from_sentence(sentence) (subject, predicate, object) = map(missing_or_resource, triple) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
<commit_before>"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() a, b, c = extract_triplet.extract_from_sentence(sentence) if a == '?': subject = ppp_datamodel.Missing() else: subject = ppp_datamodel.Resource(value=a) if b == '?': predicate = ppp_datamodel.Missing() else: predicate = ppp_datamodel.Resource(value=b) if c == '?': object = ppp_datamodel.Missing() else: object = ppp_datamodel.Resource(value=b) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]<commit_msg>Make RequestHandler's code less redundant.<commit_after>"""Request handler of the module.""" import ppp_datamodel from ppp_datamodel import Sentence, Missing, Resource from ppp_datamodel.communication import TraceItem, Response from ppp_nlp_ml_standalone import ExtractTriplet def missing_or_resource(x): return Missing() if x == '?' else Resource(value=x) class RequestHandler: def __init__(self, request): self.request = request def answer(self): if not isinstance(self.request.tree, Sentence): return [] sentence = self.request.tree.value extract_triplet = ExtractTriplet.ExtractTriplet() triple = extract_triplet.extract_from_sentence(sentence) (subject, predicate, object) = map(missing_or_resource, triple) triple = ppp_datamodel.Triple(subject=subject, predicate=predicate, object=object) meas = {'accuracy': 0.5, 'relevance': 0.5} trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)] response = Response('en', triple, meas, trace) print(repr(response)) return [response]
b4473d45ba5925551334762bc02708fcb373c957
config.py
config.py
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False }
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False, 'page_size': 100 }
Add page size for sql query
Add page size for sql query
Python
mit
goodwinxp/ATFGenerator,goodwinxp/ATFGenerator,goodwinxp/ATFGenerator
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False } Add page size for sql query
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False, 'page_size': 100 }
<commit_before>CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False } <commit_msg>Add page size for sql query<commit_after>
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False, 'page_size': 100 }
CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False } Add page size for sql queryCONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False, 'page_size': 100 }
<commit_before>CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False } <commit_msg>Add page size for sql query<commit_after>CONFIG = { 'database': './ida_info.sqlite3', 'out_dir': './code_gen/', 'verbose': False, 'page_size': 100 }
840c764dbe8da07546381b51996f64f7680327bb
sciunit/unit_test/doc_tests.py
sciunit/unit_test/doc_tests.py
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") def test_chapter6(self): self.do_notebook("chapter6")
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") #def test_chapter6(self): # self.do_notebook("chapter6")
Drop test ch. 6 since sympy is not in travis-ci
Drop test ch. 6 since sympy is not in travis-ci
Python
mit
scidash/sciunit,scidash/sciunit
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") def test_chapter6(self): self.do_notebook("chapter6") Drop test ch. 6 since sympy is not in travis-ci
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") #def test_chapter6(self): # self.do_notebook("chapter6")
<commit_before>"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") def test_chapter6(self): self.do_notebook("chapter6") <commit_msg>Drop test ch. 6 since sympy is not in travis-ci<commit_after>
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") #def test_chapter6(self): # self.do_notebook("chapter6")
"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") def test_chapter6(self): self.do_notebook("chapter6") Drop test ch. 6 since sympy is not in travis-ci"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") #def test_chapter6(self): # self.do_notebook("chapter6")
<commit_before>"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") def test_chapter6(self): self.do_notebook("chapter6") <commit_msg>Drop test ch. 6 since sympy is not in travis-ci<commit_after>"""Unit tests for documentation""" import unittest from sciunit.utils import NotebookTools class DocumentationTestCase(NotebookTools, unittest.TestCase): """Unit tests for documentation notebooks""" path = "../../docs" def test_chapter1(self): self.do_notebook("chapter1") def test_chapter2(self): self.do_notebook("chapter2") def test_chapter3(self): self.do_notebook("chapter3") def test_chapter4(self): self.do_notebook("chapter4") def test_chapter5(self): self.do_notebook("chapter5") #def test_chapter6(self): # self.do_notebook("chapter6")
cc19d0af1c22c9677960f406ced425aa48da54c1
src/sentry/migrations/0063_remove_bad_groupedmessage_index.py
src/sentry/migrations/0063_remove_bad_groupedmessage_index.py
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] try: db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) except Exception: db.rollback_transaction() def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
Revert "Dont error if 0063 index was already cleaned up"
Revert "Dont error if 0063 index was already cleaned up" This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.
Python
bsd-3-clause
felixbuenemann/sentry,JackDanger/sentry,zenefits/sentry,korealerts1/sentry,fuziontech/sentry,daevaorn/sentry,argonemyth/sentry,beeftornado/sentry,vperron/sentry,mvaled/sentry,rdio/sentry,gg7/sentry,hongliang5623/sentry,felixbuenemann/sentry,jokey2k/sentry,pauloschilling/sentry,beni55/sentry,rdio/sentry,BayanGroup/sentry,ngonzalvez/sentry,gencer/sentry,mitsuhiko/sentry,Natim/sentry,gg7/sentry,boneyao/sentry,argonemyth/sentry,NickPresta/sentry,llonchj/sentry,SilentCircle/sentry,looker/sentry,Kryz/sentry,ngonzalvez/sentry,daevaorn/sentry,jokey2k/sentry,alexm92/sentry,BuildingLink/sentry,JamesMura/sentry,pauloschilling/sentry,drcapulet/sentry,beni55/sentry,camilonova/sentry,looker/sentry,vperron/sentry,JamesMura/sentry,mvaled/sentry,songyi199111/sentry,fotinakis/sentry,1tush/sentry,gencer/sentry,jean/sentry,wujuguang/sentry,ifduyue/sentry,JTCunning/sentry,fuziontech/sentry,vperron/sentry,mvaled/sentry,fuziontech/sentry,beeftornado/sentry,Kryz/sentry,camilonova/sentry,ewdurbin/sentry,kevinastone/sentry,llonchj/sentry,gencer/sentry,nicholasserra/sentry,jean/sentry,TedaLIEz/sentry,drcapulet/sentry,mvaled/sentry,looker/sentry,imankulov/sentry,gg7/sentry,NickPresta/sentry,songyi199111/sentry,mvaled/sentry,SilentCircle/sentry,JamesMura/sentry,BuildingLink/sentry,kevinlondon/sentry,Natim/sentry,hongliang5623/sentry,wujuguang/sentry,felixbuenemann/sentry,kevinastone/sentry,gencer/sentry,daevaorn/sentry,JamesMura/sentry,korealerts1/sentry,korealerts1/sentry,alexm92/sentry,BuildingLink/sentry,ifduyue/sentry,looker/sentry,beeftornado/sentry,JackDanger/sentry,BayanGroup/sentry,gencer/sentry,fotinakis/sentry,songyi199111/sentry,beni55/sentry,nicholasserra/sentry,Natim/sentry,zenefits/sentry,hongliang5623/sentry,1tush/sentry,argonemyth/sentry,llonchj/sentry,alexm92/sentry,zenefits/sentry,JackDanger/sentry,ifduyue/sentry,1tush/sentry,boneyao/sentry,JamesMura/sentry,SilentCircle/sentry,pauloschilling/sentry,ewdurbin/sentry,zenefits/sentry,mvaled/sentry,imankulov/sentry,mitsuhiko/sentry,Kryz/sentry,looker/sentry,kevinastone/sentry,kevinlondon/sentry,zenefits/sentry,ifduyue/sentry,NickPresta/sentry,wong2/sentry,ngonzalvez/sentry,rdio/sentry,fotinakis/sentry,jokey2k/sentry,SilentCircle/sentry,TedaLIEz/sentry,fotinakis/sentry,nicholasserra/sentry,daevaorn/sentry,TedaLIEz/sentry,camilonova/sentry,wong2/sentry,BuildingLink/sentry,jean/sentry,JTCunning/sentry,imankulov/sentry,jean/sentry,ifduyue/sentry,JTCunning/sentry,NickPresta/sentry,kevinlondon/sentry,drcapulet/sentry,rdio/sentry,wujuguang/sentry,jean/sentry,BayanGroup/sentry,boneyao/sentry,ewdurbin/sentry,wong2/sentry,BuildingLink/sentry
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] try: db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) except Exception: db.rollback_transaction() def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) Revert "Dont error if 0063 index was already cleaned up" This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
<commit_before># encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] try: db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) except Exception: db.rollback_transaction() def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) <commit_msg>Revert "Dont error if 0063 index was already cleaned up" This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.<commit_after>
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] try: db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) except Exception: db.rollback_transaction() def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) Revert "Dont error if 0063 index was already cleaned up" This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.# encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
<commit_before># encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] try: db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) except Exception: db.rollback_transaction() def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) <commit_msg>Revert "Dont error if 0063 index was already cleaned up" This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.<commit_after># encoding: utf-8 import datetime from south.db import db from south.v2 import SchemaMigration from django.db import models class Migration(SchemaMigration): def forwards(self, orm): # Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum']) def backwards(self, orm): # Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum'] db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
da0d5d5058e57b3f2dd2b16822e0ab47ced7353a
AddressServer.py
AddressServer.py
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose SocketServer.TCPServer.allow_reuse_address = True server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
Allow re-use of a port a la SO_REUSEADDRESS
Allow re-use of a port a la SO_REUSEADDRESS
Python
apache-2.0
mattrichards/TCPAddressServer
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main() Allow re-use of a port a la SO_REUSEADDRESS
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose SocketServer.TCPServer.allow_reuse_address = True server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
<commit_before>#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main() <commit_msg>Allow re-use of a port a la SO_REUSEADDRESS<commit_after>
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose SocketServer.TCPServer.allow_reuse_address = True server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main() Allow re-use of a port a la SO_REUSEADDRESS#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose SocketServer.TCPServer.allow_reuse_address = True server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
<commit_before>#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main() <commit_msg>Allow re-use of a port a la SO_REUSEADDRESS<commit_after>#!/usr/bin/python """ Simple TCP server that sends back the IP address and port from which it sees you connect. Useful when used with `telnet` for network debugging. Currently only IPv4. """ import SocketServer VERBOSE = False class TCPHandler(SocketServer.BaseRequestHandler): """ Handles each request be sending the client's IP address and port, then closing the connection. """ def handle(self): # self.request is the TCP socket connected to the client address = self.client_address[0] + ":" + str(self.client_address[1]) if VERBOSE: print "Request from: " + address self.request.sendall(address + "\n") def main(): import optparse global VERBOSE parser = optparse.OptionParser() parser.add_option("-p", "--port", type="int", default=9999, help="port to listen on") parser.add_option("-v", "--verbose", action="store_true") options = parser.parse_args()[0] port = options.port VERBOSE = options.verbose SocketServer.TCPServer.allow_reuse_address = True server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler) print "Listening on port %d" % port # Activate the server; this will keep running until you # interrupt the program with Ctrl-C server.serve_forever() if __name__ == "__main__": main()
1daeedde2cd8597e047b6a6d7fc737f103fa4ac8
example/handler/my_handler.py
example/handler/my_handler.py
from base_handler import BaseHandler from utils import truncated_stdout, with_payload class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): print("Hello there, {}!".format(who))
from base_handler import BaseHandler from utils import truncated_stdout, with_payload, with_member_info class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): """A custom user event.""" print("Hello there, {}!".format(who)) @with_payload def supervisor(self, **kwargs): """This event gets fired for change of state in supervisor.""" print("Got a supervisor event with payload:") print(kwargs) @with_member_info def member_join(self, members): """This event gets fired on members joining. ``members`` is the dict with new members' """ print("Hello to new members:") print(members)
Add example of supervisor event and member join
Add example of supervisor event and member join
Python
mit
waltermoreira/serfnode,waltermoreira/serfnode,waltermoreira/serfnode
from base_handler import BaseHandler from utils import truncated_stdout, with_payload class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): print("Hello there, {}!".format(who)) Add example of supervisor event and member join
from base_handler import BaseHandler from utils import truncated_stdout, with_payload, with_member_info class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): """A custom user event.""" print("Hello there, {}!".format(who)) @with_payload def supervisor(self, **kwargs): """This event gets fired for change of state in supervisor.""" print("Got a supervisor event with payload:") print(kwargs) @with_member_info def member_join(self, members): """This event gets fired on members joining. ``members`` is the dict with new members' """ print("Hello to new members:") print(members)
<commit_before>from base_handler import BaseHandler from utils import truncated_stdout, with_payload class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): print("Hello there, {}!".format(who)) <commit_msg>Add example of supervisor event and member join<commit_after>
from base_handler import BaseHandler from utils import truncated_stdout, with_payload, with_member_info class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): """A custom user event.""" print("Hello there, {}!".format(who)) @with_payload def supervisor(self, **kwargs): """This event gets fired for change of state in supervisor.""" print("Got a supervisor event with payload:") print(kwargs) @with_member_info def member_join(self, members): """This event gets fired on members joining. ``members`` is the dict with new members' """ print("Hello to new members:") print(members)
from base_handler import BaseHandler from utils import truncated_stdout, with_payload class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): print("Hello there, {}!".format(who)) Add example of supervisor event and member joinfrom base_handler import BaseHandler from utils import truncated_stdout, with_payload, with_member_info class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): """A custom user event.""" print("Hello there, {}!".format(who)) @with_payload def supervisor(self, **kwargs): """This event gets fired for change of state in supervisor.""" print("Got a supervisor event with payload:") print(kwargs) @with_member_info def member_join(self, members): """This event gets fired on members joining. ``members`` is the dict with new members' """ print("Hello to new members:") print(members)
<commit_before>from base_handler import BaseHandler from utils import truncated_stdout, with_payload class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): print("Hello there, {}!".format(who)) <commit_msg>Add example of supervisor event and member join<commit_after>from base_handler import BaseHandler from utils import truncated_stdout, with_payload, with_member_info class MyHandler(BaseHandler): @truncated_stdout @with_payload def hello(self, who=None): """A custom user event.""" print("Hello there, {}!".format(who)) @with_payload def supervisor(self, **kwargs): """This event gets fired for change of state in supervisor.""" print("Got a supervisor event with payload:") print(kwargs) @with_member_info def member_join(self, members): """This event gets fired on members joining. ``members`` is the dict with new members' """ print("Hello to new members:") print(members)
54d4551ce8efb16d4a8d02e38b9f223f8f1cd816
ab_game.py
ab_game.py
#!/usr/bin/python import board import pente_exceptions from ab_state import * CAPTURE_SCORE_BASE = 120 ** 3 class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 if depth > 4: min_priority = 3 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
#!/usr/bin/python import board import pente_exceptions from ab_state import * class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
Disable min_priority filter for now
Disable min_priority filter for now
Python
mit
cropleyb/pentai,cropleyb/pentai,cropleyb/pentai
#!/usr/bin/python import board import pente_exceptions from ab_state import * CAPTURE_SCORE_BASE = 120 ** 3 class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 if depth > 4: min_priority = 3 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal() Disable min_priority filter for now
#!/usr/bin/python import board import pente_exceptions from ab_state import * class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
<commit_before>#!/usr/bin/python import board import pente_exceptions from ab_state import * CAPTURE_SCORE_BASE = 120 ** 3 class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 if depth > 4: min_priority = 3 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal() <commit_msg>Disable min_priority filter for now<commit_after>
#!/usr/bin/python import board import pente_exceptions from ab_state import * class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
#!/usr/bin/python import board import pente_exceptions from ab_state import * CAPTURE_SCORE_BASE = 120 ** 3 class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 if depth > 4: min_priority = 3 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal() Disable min_priority filter for now#!/usr/bin/python import board import pente_exceptions from ab_state import * class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
<commit_before>#!/usr/bin/python import board import pente_exceptions from ab_state import * CAPTURE_SCORE_BASE = 120 ** 3 class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 if depth > 4: min_priority = 3 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal() <commit_msg>Disable min_priority filter for now<commit_after>#!/usr/bin/python import board import pente_exceptions from ab_state import * class ABGame(): """ This class acts as a bridge between the AlphaBeta code and my code """ def __init__(self, base_game): s = self.current_state = ABState() s.set_state(base_game.current_state) self.base_game = base_game def to_move(self, state=None): if state is None: state = self.current_state return state.to_move() def utility(self, state): return state.utility() def successors(self, state, depth): mn = state.get_move_number() if mn == 1: # The first black move is always in the centre brd_size = self.base_game.get_board().get_size() centre_pos = (brd_size/2, brd_size/2) p_i = [centre_pos] else: min_priority = 0 pos_iter = state.get_iter(state.to_move()) p_i = pos_iter.get_iter(state.to_move_colour(), min_priority) tried_count = 0 for pos in p_i: # create an AB_State for each possible move from state succ = state.create_state(pos) yield pos, succ tried_count += 1 if depth > 3 and tried_count >= 2: return def terminal_test(self, state): return state.terminal()
06d9171b2244e4dd9d5e1883101d7ec3e05be4b2
bitfield/apps.py
bitfield/apps.py
from django.apps import AppConfig class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
import django from django.apps import AppConfig django.setup() class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
Add django.setup to the AppConfig
Add django.setup to the AppConfig
Python
apache-2.0
Elec/django-bitfield,disqus/django-bitfield,joshowen/django-bitfield
from django.apps import AppConfig class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field" Add django.setup to the AppConfig
import django from django.apps import AppConfig django.setup() class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
<commit_before>from django.apps import AppConfig class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field" <commit_msg>Add django.setup to the AppConfig<commit_after>
import django from django.apps import AppConfig django.setup() class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
from django.apps import AppConfig class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field" Add django.setup to the AppConfigimport django from django.apps import AppConfig django.setup() class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
<commit_before>from django.apps import AppConfig class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field" <commit_msg>Add django.setup to the AppConfig<commit_after>import django from django.apps import AppConfig django.setup() class BitFieldAppConfig(AppConfig): name = 'bitfield' verbose_name = "Bit Field"
0db43d894bfb419a7f4b538f755af47fc0b653cb
tests/unit/test_sharpspring.py
tests/unit/test_sharpspring.py
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called()
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
Add mock response to sharpspring test
Add mock response to sharpspring test
Python
apache-2.0
Code4SA/pmg-cms-2,Code4SA/pmg-cms-2,Code4SA/pmg-cms-2
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() Add mock response to sharpspring test
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
<commit_before>from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() <commit_msg>Add mock response to sharpspring test<commit_after>
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() Add mock response to sharpspring testfrom unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
<commit_before>from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post") def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() <commit_msg>Add mock response to sharpspring test<commit_after>from unittest.mock import patch from pmg.sharpspring import Sharpspring from tests import PMGTestCase class MockResponse: def __init__(self, json_data, status_code): self.json_data = json_data self.status_code = status_code def raise_for_status(self): pass def json(self): return self.json_data def mocked_requests_post_success(*args, **kwargs): return MockResponse({"result": {"creates": [{"success": True,}]}}, 200) class TestSharpspring(PMGTestCase): @patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success) def test_make_sharpsrping_request(self, post_mock): sharpspring = Sharpspring() details = { "emailAddress": "test@example.com", "companyName": "Test Company", } result = sharpspring.call("createLeads", {"objects": [details]}) post_mock.assert_called() self.assertTrue(result["result"]["creates"][0])
0cb807470ee56207251f36ad78d35c48f6e9361b
example_project/urls.py
example_project/urls.py
from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
from django.conf.urls import include, url from django.contrib import admin admin.autodiscover() # For Django 1.6 urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
Python
mit
BocuStudio/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,BocuStudio/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,BocuStudio/django-timepiece
from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ] Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
from django.conf.urls import include, url from django.contrib import admin admin.autodiscover() # For Django 1.6 urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
<commit_before>from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ] <commit_msg>Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility<commit_after>
from django.conf.urls import include, url from django.contrib import admin admin.autodiscover() # For Django 1.6 urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ] Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibilityfrom django.conf.urls import include, url from django.contrib import admin admin.autodiscover() # For Django 1.6 urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
<commit_before>from django.conf.urls import include, url from django.contrib import admin urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ] <commit_msg>Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility<commit_after>from django.conf.urls import include, url from django.contrib import admin admin.autodiscover() # For Django 1.6 urlpatterns = [ url(r'^admin/', include(admin.site.urls)), url(r'^selectable/', include('selectable.urls')), url(r'', include('timepiece.urls')), # authentication views url(r'^accounts/login/$', 'django.contrib.auth.views.login', name='auth_login'), url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login', name='auth_logout'), url(r'^accounts/password-change/$', 'django.contrib.auth.views.password_change', name='change_password'), url(r'^accounts/password-change/done/$', 'django.contrib.auth.views.password_change_done'), url(r'^accounts/password-reset/$', 'django.contrib.auth.views.password_reset', name='reset_password'), url(r'^accounts/password-reset/done/$', 'django.contrib.auth.views.password_reset_done'), url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$', 'django.contrib.auth.views.password_reset_confirm'), url(r'^accounts/reset/done/$', 'django.contrib.auth.views.password_reset_complete'), ]
23343eb3316a3d304a3b021519b9a470f9c2446b
django_bcrypt/models.py
django_bcrypt/models.py
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): if raw_password is None: self.set_unusable_password() else: salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
Allow users to be created with blank (unusable) passwords.
Allow users to be created with blank (unusable) passwords.
Python
mit
dwaiter/django-bcrypt
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password Allow users to be created with blank (unusable) passwords.
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): if raw_password is None: self.set_unusable_password() else: salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
<commit_before>import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password <commit_msg>Allow users to be created with blank (unusable) passwords.<commit_after>
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): if raw_password is None: self.set_unusable_password() else: salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password Allow users to be created with blank (unusable) passwords.import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): if raw_password is None: self.set_unusable_password() else: salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
<commit_before>import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password <commit_msg>Allow users to be created with blank (unusable) passwords.<commit_after>import bcrypt from django.contrib.auth.models import User from django.conf import settings try: rounds = settings.BCRYPT_ROUNDS except AttributeError: rounds = 12 _check_password = User.check_password def bcrypt_check_password(self, raw_password): if self.password.startswith('bc$'): salt_and_hash = self.password[3:] return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash return _check_password(self, raw_password) def bcrypt_set_password(self, raw_password): if raw_password is None: self.set_unusable_password() else: salt = bcrypt.gensalt(rounds) self.password = 'bc$' + bcrypt.hashpw(raw_password, salt) User.check_password = bcrypt_check_password User.set_password = bcrypt_set_password
f5e4a8000e23e279192834d03e4b5b9ecca6b2b0
linguist/utils/__init__.py
linguist/utils/__init__.py
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, get_real_field_name, get_fallback_field_name, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
Fix new i18n utils imports.
Fix new i18n utils imports.
Python
mit
ulule/django-linguist
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n] Fix new i18n utils imports.
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, get_real_field_name, get_fallback_field_name, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
<commit_before># -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n] <commit_msg>Fix new i18n utils imports.<commit_after>
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, get_real_field_name, get_fallback_field_name, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n] Fix new i18n utils imports.# -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, get_real_field_name, get_fallback_field_name, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
<commit_before># -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n] <commit_msg>Fix new i18n utils imports.<commit_after># -*- coding: utf-8 -*- from .i18n import (get_language_name, get_language, get_fallback_language, get_real_field_name, get_fallback_field_name, build_localized_field_name, build_localized_verbose_name) from .models import load_class, get_model_string from .template import select_template_name from .views import get_language_parameter, get_language_tabs __all__ = [ 'get_language_name', 'get_language', 'get_fallback_language', 'build_localized_field_name', 'build_localized_verbose_name', 'load_class', 'get_model_string', 'select_template_name', 'get_language_parameter', 'get_language_tabs', 'chunks', ] def chunks(l, n): """ Yields successive n-sized chunks from l. """ for i in xrange(0, len(l), n): yield l[i:i + n]
986b9227fe66d95a7e42253395c89de5c2385b2d
scuole/campuses/management/commands/dedupecampusslugs.py
scuole/campuses/management/commands/dedupecampusslugs.py
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) print('DUPLICATE SLUGS', duplicate_slugs) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] # for campus in Campus.objects.filter(slug=slug): # if campus.city != None: # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # campus.save() # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # print(slugify(campus.city, allow_unicode=True)) # print('SLUG', campus.slug) # campus.save()
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] campuses_dup_slug = Campus.objects.filter(slug=slug) # if the district and county are the same, but the city of the campuses are different if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug): for campus in campuses_dup_slug: if campus.city != None: city_slug = slugify(campus.city, allow_unicode=True) campus.slug = f"{campus.slug}-{city_slug}" campus.save() # if the district, county, and city of the campuses are the same if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug): for campus in campuses_dup_slug: campus.slug = f"{campus.slug}-{campus.tea_id}" campus.save()
Edit dedupe campus slugs code
Edit dedupe campus slugs code
Python
mit
texastribune/scuole,texastribune/scuole,texastribune/scuole,texastribune/scuole
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) print('DUPLICATE SLUGS', duplicate_slugs) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] # for campus in Campus.objects.filter(slug=slug): # if campus.city != None: # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # campus.save() # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # print(slugify(campus.city, allow_unicode=True)) # print('SLUG', campus.slug) # campus.save() Edit dedupe campus slugs code
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] campuses_dup_slug = Campus.objects.filter(slug=slug) # if the district and county are the same, but the city of the campuses are different if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug): for campus in campuses_dup_slug: if campus.city != None: city_slug = slugify(campus.city, allow_unicode=True) campus.slug = f"{campus.slug}-{city_slug}" campus.save() # if the district, county, and city of the campuses are the same if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug): for campus in campuses_dup_slug: campus.slug = f"{campus.slug}-{campus.tea_id}" campus.save()
<commit_before>from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) print('DUPLICATE SLUGS', duplicate_slugs) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] # for campus in Campus.objects.filter(slug=slug): # if campus.city != None: # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # campus.save() # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # print(slugify(campus.city, allow_unicode=True)) # print('SLUG', campus.slug) # campus.save() <commit_msg>Edit dedupe campus slugs code<commit_after>
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] campuses_dup_slug = Campus.objects.filter(slug=slug) # if the district and county are the same, but the city of the campuses are different if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug): for campus in campuses_dup_slug: if campus.city != None: city_slug = slugify(campus.city, allow_unicode=True) campus.slug = f"{campus.slug}-{city_slug}" campus.save() # if the district, county, and city of the campuses are the same if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug): for campus in campuses_dup_slug: campus.slug = f"{campus.slug}-{campus.tea_id}" campus.save()
from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) print('DUPLICATE SLUGS', duplicate_slugs) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] # for campus in Campus.objects.filter(slug=slug): # if campus.city != None: # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # campus.save() # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # print(slugify(campus.city, allow_unicode=True)) # print('SLUG', campus.slug) # campus.save() Edit dedupe campus slugs codefrom django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] campuses_dup_slug = Campus.objects.filter(slug=slug) # if the district and county are the same, but the city of the campuses are different if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug): for campus in campuses_dup_slug: if campus.city != None: city_slug = slugify(campus.city, allow_unicode=True) campus.slug = f"{campus.slug}-{city_slug}" campus.save() # if the district, county, and city of the campuses are the same if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug): for campus in campuses_dup_slug: campus.slug = f"{campus.slug}-{campus.tea_id}" campus.save()
<commit_before>from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) print('DUPLICATE SLUGS', duplicate_slugs) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] # for campus in Campus.objects.filter(slug=slug): # if campus.city != None: # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # campus.save() # city_slug = slugify(campus.city, allow_unicode=True) # campus.slug = f"{campus.slug}-{city_slug}" # print(slugify(campus.city, allow_unicode=True)) # print('SLUG', campus.slug) # campus.save() <commit_msg>Edit dedupe campus slugs code<commit_after>from django.core.management.base import BaseCommand from django.db.models import Count from django.utils.text import slugify from scuole.campuses.models import Campus class Command(BaseCommand): help = "Dedupe Campus slugs by adding the county name to the end." def handle(self, *args, **options): duplicate_slugs = ( Campus.objects.values("slug") .annotate(total=Count("slug")) .filter(total__gt=1) ) # loop through all duplicate slugs for duplicate in duplicate_slugs: slug = duplicate['slug'] campuses_dup_slug = Campus.objects.filter(slug=slug) # if the district and county are the same, but the city of the campuses are different if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug): for campus in campuses_dup_slug: if campus.city != None: city_slug = slugify(campus.city, allow_unicode=True) campus.slug = f"{campus.slug}-{city_slug}" campus.save() # if the district, county, and city of the campuses are the same if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug): for campus in campuses_dup_slug: campus.slug = f"{campus.slug}-{campus.tea_id}" campus.save()
de0bb4886b9a6ecd2fb4e5c4272167911141c71c
apic_ml2/neutron/plugins/ml2/drivers/cisco/apic/nova_client.py
apic_ml2/neutron/plugins/ml2/drivers/cisco/apic/nova_client.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) class NovaClient(object): def __init__(self): self.nclient = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.nclient.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) client = None def _get_client(): global client if client is None: client = n_nova.Notifier().nclient return client class NovaClient(object): def __init__(self): self.client = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.client.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
Load Nova Client only once to avoid reconnecting
Load Nova Client only once to avoid reconnecting
Python
apache-2.0
noironetworks/apic-ml2-driver,noironetworks/apic-ml2-driver
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) class NovaClient(object): def __init__(self): self.nclient = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.nclient.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e) Load Nova Client only once to avoid reconnecting
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) client = None def _get_client(): global client if client is None: client = n_nova.Notifier().nclient return client class NovaClient(object): def __init__(self): self.client = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.client.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) class NovaClient(object): def __init__(self): self.nclient = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.nclient.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e) <commit_msg>Load Nova Client only once to avoid reconnecting<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) client = None def _get_client(): global client if client is None: client = n_nova.Notifier().nclient return client class NovaClient(object): def __init__(self): self.client = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.client.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) class NovaClient(object): def __init__(self): self.nclient = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.nclient.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e) Load Nova Client only once to avoid reconnecting# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) client = None def _get_client(): global client if client is None: client = n_nova.Notifier().nclient return client class NovaClient(object): def __init__(self): self.client = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.client.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) class NovaClient(object): def __init__(self): self.nclient = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.nclient.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e) <commit_msg>Load Nova Client only once to avoid reconnecting<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from neutron._i18n import _LW from neutron.notifiers import nova as n_nova from novaclient import exceptions as nova_exceptions from oslo_log import log as logging LOG = logging.getLogger(__name__) client = None def _get_client(): global client if client is None: client = n_nova.Notifier().nclient return client class NovaClient(object): def __init__(self): self.client = n_nova.Notifier().nclient def get_server(self, server_id): try: return self.client.servers.get(server_id) except nova_exceptions.NotFound: LOG.warning(_LW("Nova returned NotFound for server: %s"), server_id) except Exception as e: LOG.exception(e)
79c5a3b12fbe0ccde4bf8ec8694d42696241621d
products/bika/browser/clientfolder.py
products/bika/browser/clientfolder.py
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'}, 'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'}, 'getPhone': {'title': 'Phone', 'field':'getPhone'}, 'getFax': {'title': 'Fax', 'field':'getFax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name'}, 'getEmailAddress': {'title': 'Email Address'}, 'getPhone': {'title': 'Phone'}, 'getFax': {'title': 'Fax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()
Remove 'field' and 'icon' from column list
Remove 'field' and 'icon' from column list
Python
agpl-3.0
veroc/Bika-LIMS,veroc/Bika-LIMS,veroc/Bika-LIMS,DeBortoliWines/Bika-LIMS,labsanmartin/Bika-LIMS,rockfruit/bika.lims,labsanmartin/Bika-LIMS,DeBortoliWines/Bika-LIMS,labsanmartin/Bika-LIMS,anneline/Bika-LIMS,anneline/Bika-LIMS,DeBortoliWines/Bika-LIMS,anneline/Bika-LIMS,rockfruit/bika.lims
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'}, 'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'}, 'getPhone': {'title': 'Phone', 'field':'getPhone'}, 'getFax': {'title': 'Fax', 'field':'getFax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template() Remove 'field' and 'icon' from column list
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name'}, 'getEmailAddress': {'title': 'Email Address'}, 'getPhone': {'title': 'Phone'}, 'getFax': {'title': 'Fax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()
<commit_before>from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'}, 'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'}, 'getPhone': {'title': 'Phone', 'field':'getPhone'}, 'getFax': {'title': 'Fax', 'field':'getFax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template() <commit_msg>Remove 'field' and 'icon' from column list<commit_after>
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name'}, 'getEmailAddress': {'title': 'Email Address'}, 'getPhone': {'title': 'Phone'}, 'getFax': {'title': 'Fax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()
from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'}, 'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'}, 'getPhone': {'title': 'Phone', 'field':'getPhone'}, 'getFax': {'title': 'Fax', 'field':'getFax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template() Remove 'field' and 'icon' from column listfrom Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name'}, 'getEmailAddress': {'title': 'Email Address'}, 'getPhone': {'title': 'Phone'}, 'getFax': {'title': 'Fax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()
<commit_before>from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'}, 'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'}, 'getPhone': {'title': 'Phone', 'field':'getPhone'}, 'getFax': {'title': 'Fax', 'field':'getFax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template() <commit_msg>Remove 'field' and 'icon' from column list<commit_after>from Products.CMFCore.utils import getToolByName from Products.bika import logger from Products.bika.browser.bika_folder_contents import BikaFolderContentsView from plone.app.content.browser.interfaces import IFolderContentsView from zope.interface import implements class ClientFolderContentsView(BikaFolderContentsView): implements(IFolderContentsView) contentFilter = {'portal_type': 'Client'} content_add_buttons = ['Client', ] batch = True b_size = 100 show_editable_border = False columns = { 'title_or_id': {'title': 'Name'}, 'getEmailAddress': {'title': 'Email Address'}, 'getPhone': {'title': 'Phone'}, 'getFax': {'title': 'Fax'}, } wflist_states = [ {'title': 'All', 'id':'all', 'columns':['title_or_id', 'getEmailAddress', 'getPhone', 'getFax', ], 'buttons':[BikaFolderContentsView.default_buttons['delete']]}, ] def folderitems(self): items = BikaFolderContentsView.folderitems(self) for x in range(len(items)): items[x]['links'] = {'title_or_id': items[x]['url']} return items def __call__(self): return self.template()