commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 0
2.94k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
444
| message
stringlengths 16
3.45k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43.2k
| prompt
stringlengths 17
4.58k
| response
stringlengths 1
4.43k
| prompt_tagged
stringlengths 58
4.62k
| response_tagged
stringlengths 1
4.43k
| text
stringlengths 132
7.29k
| text_tagged
stringlengths 173
7.33k
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
82e964dab398caee75c3174f86593ab6cfa7dbaf
|
src/constants.py
|
src/constants.py
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 40.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 80.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
Increase simulation time for linear trajectory
|
Increase simulation time for linear trajectory
|
Python
|
mit
|
bit0001/trajectory_tracking,bit0001/trajectory_tracking
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 40.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
Increase simulation time for linear trajectory
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 80.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
<commit_before>#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 40.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
<commit_msg>Increase simulation time for linear trajectory<commit_after>
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 80.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 40.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
Increase simulation time for linear trajectory#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 80.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
<commit_before>#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 40.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
<commit_msg>Increase simulation time for linear trajectory<commit_after>#!/usr/bin/env python
TRAJECTORY = 'linear'
if TRAJECTORY == 'linear':
SIMULATION_TIME_IN_SECONDS = 80.0
elif TRAJECTORY == 'circular':
SIMULATION_TIME_IN_SECONDS = 120.0
elif TRAJECTORY == 'squared':
SIMULATION_TIME_IN_SECONDS = 160.0
DELTA_T = 0.1 # this is the sampling time
STEPS = int(SIMULATION_TIME_IN_SECONDS / DELTA_T)
# control constants
K_X = 0.90
K_Y = 0.90
K_THETA = 0.90
# PID control constants
K_P_V = 0.2
K_I_V = 1.905
K_D_V = 0.00
K_P_W = 0.45
K_I_W = 1.25
K_D_W = 0.000
|
63e61d9123a4abeb85d83545da351ed0abf792aa
|
src/core/urls.py
|
src/core/urls.py
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
try:
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
except AttributeError:
pass
|
Handle installs not using new settings engine
|
Handle installs not using new settings engine
|
Python
|
agpl-3.0
|
BirkbeckCTP/janeway,BirkbeckCTP/janeway,BirkbeckCTP/janeway,BirkbeckCTP/janeway
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
Handle installs not using new settings engine
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
try:
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
except AttributeError:
pass
|
<commit_before>__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
<commit_msg>Handle installs not using new settings engine<commit_after>
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
try:
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
except AttributeError:
pass
|
__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
Handle installs not using new settings engine__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
try:
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
except AttributeError:
pass
|
<commit_before>__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
<commit_msg>Handle installs not using new settings engine<commit_after>__copyright__ = "Copyright 2017 Birkbeck, University of London"
__author__ = "Martin Paul Eve & Andy Byers"
__license__ = "AGPL v3"
__maintainer__ = "Birkbeck Centre for Technology and Publishing"
from django.conf.urls import include, url
from django.contrib import admin
from django.views.generic import TemplateView
from django.conf import settings
from django.views.static import serve
from press import views as press_views
include('events.registration')
urlpatterns = [
url(r'^$', press_views.index, name='website_index'),
url(r'^admin/', include(admin.site.urls)),
url(r'^summernote/', include('django_summernote.urls')),
url(r'', include('core.include_urls')),
]
try:
if settings.DEBUG or settings.IN_TEST_RUNNER:
import debug_toolbar
urlpatterns += [
url(r'^media/(?P<path>.*)$', serve, {'document_root': settings.MEDIA_ROOT}),
url(r'^404/$', TemplateView.as_view(template_name='404.html')),
url(r'^500/$', TemplateView.as_view(template_name='500.html')),
url(r'^__debug__/', include(debug_toolbar.urls)),
url(r'^hijack/', include('hijack.urls', namespace='hijack')),
]
except AttributeError:
pass
|
1a5c1ea0815d30048d7dbce56adf2503b9c82c28
|
moksha/widgets/container/tests/test_container.py
|
moksha/widgets/container/tests/test_container.py
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert '<div id="test" ' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert 'Moksha Container' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
Update our container test case
|
Update our container test case
|
Python
|
apache-2.0
|
ralphbean/moksha,pombredanne/moksha,pombredanne/moksha,lmacken/moksha,mokshaproject/moksha,mokshaproject/moksha,lmacken/moksha,mokshaproject/moksha,pombredanne/moksha,mokshaproject/moksha,ralphbean/moksha,ralphbean/moksha,lmacken/moksha,pombredanne/moksha
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert '<div id="test" ' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
Update our container test case
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert 'Moksha Container' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
<commit_before># This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert '<div id="test" ' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
<commit_msg>Update our container test case<commit_after>
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert 'Moksha Container' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert '<div id="test" ' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
Update our container test case# This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert 'Moksha Container' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
<commit_before># This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert '<div id="test" ' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
<commit_msg>Update our container test case<commit_after># This file is part of Moksha.
# Copyright (C) 2008-2009 Red Hat, Inc.
#
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as
# published by the Free Software Foundation, either version 3 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
from tw.api import Widget
from moksha.widgets.container import MokshaContainer
class TestContainer:
def setUp(self):
self.w = MokshaContainer('test')
def test_render_widget(self):
assert 'Moksha Container' in self.w()
def test_widget_content(self):
""" Ensure we can render a container with another widget """
class MyWidget(Widget):
template = """
Hello World!
"""
assert 'Hello World!' in self.w(content=MyWidget('mywidget'))
def test_container_classes(self):
rendered = self.w(**dict(skin3=True, stikynote=True,
draggable=True, resizable=True))
assert 'class="containerPlus draggable resizable"' in rendered, rendered
|
35ffe6bb97a30970d4bc3c265b6337712669ee09
|
githubsetupircnotifications.py
|
githubsetupircnotifications.py
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.organization)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def error(message):
print(message)
sys.exit(1)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
error('Failed to sign into github')
org = github.organization(args.organization)
if org is None:
error('Organization "{}" does not appear to exist'.format(args.org))
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
Add error function to remove duplicate code
|
Add error function to remove duplicate code
|
Python
|
mit
|
kragniz/github-setup-irc-notifications
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.organization)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
Add error function to remove duplicate code
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def error(message):
print(message)
sys.exit(1)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
error('Failed to sign into github')
org = github.organization(args.organization)
if org is None:
error('Organization "{}" does not appear to exist'.format(args.org))
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
<commit_before>"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.organization)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
<commit_msg>Add error function to remove duplicate code<commit_after>
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def error(message):
print(message)
sys.exit(1)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
error('Failed to sign into github')
org = github.organization(args.organization)
if org is None:
error('Organization "{}" does not appear to exist'.format(args.org))
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.organization)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
Add error function to remove duplicate code"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def error(message):
print(message)
sys.exit(1)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
error('Failed to sign into github')
org = github.organization(args.organization)
if org is None:
error('Organization "{}" does not appear to exist'.format(args.org))
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
<commit_before>"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
print('Failed to sign into github')
sys.exit(1)
org = github.organization(args.organization)
if org is None:
print('Organization "{}" does not appear to exist'.format(args.org))
sys.exit(1)
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
<commit_msg>Add error function to remove duplicate code<commit_after>"""
github-setup-irc-notifications - Configure all repositories in an organization
with irc notifications
"""
import argparse
import getpass
import sys
import github3
def error(message):
print(message)
sys.exit(1)
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--username')
parser.add_argument('--password')
parser.add_argument('organization')
parser.add_argument('channel')
args = parser.parse_args()
if args.password is None:
password = getpass.getpass(
'Password for github user "{}":'.format(args.username))
else:
password = args.password
github = github3.login(args.username, password=password)
if github is None:
error('Failed to sign into github')
org = github.organization(args.organization)
if org is None:
error('Organization "{}" does not appear to exist'.format(args.org))
conf = {'nickserv_password': '',
'no_colors': '0',
'password': '',
'branch_regexes': '',
'room': args.channel,
'ssl': '0',
'port': '',
'branches': '',
'server': 'chat.freenode.net',
'long_url': '0',
'notice': '0',
'message_without_join': '1',
'nick': 'github'
}
events = [
'push',
'delete',
'create',
'issues',
'pull_request'
]
for r in org.iter_repos():
r.create_hook('irc', conf, events=events)
|
859a23790968c84cdbc4fa7467957a3a1ed1e069
|
greatbigcrane/project/forms.py
|
greatbigcrane/project/forms.py
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
name = forms.CharField(initial="django")
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
Add a name for the recipe section
|
Add a name for the recipe section
|
Python
|
apache-2.0
|
pnomolos/greatbigcrane,pnomolos/greatbigcrane
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
Add a name for the recipe section
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
name = forms.CharField(initial="django")
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
<commit_before>"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
<commit_msg>Add a name for the recipe section<commit_after>
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
name = forms.CharField(initial="django")
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
Add a name for the recipe section"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
name = forms.CharField(initial="django")
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
<commit_before>"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
<commit_msg>Add a name for the recipe section<commit_after>"""
Copyright 2010 Jason Chu, Dusty Phillips, and Phil Schalm
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
from django import forms
from project.models import Project
class ProjectForm(forms.ModelForm):
class Meta:
model = Project
class DjangoRecipeForm(forms.Form):
name = forms.CharField(initial="django")
settings = forms.CharField()
version = forms.ChoiceField(choices=[
("trunk", "trunk"),
("1.2", "1.2"),
("1.1.2", "1.1.2"),
("1.1", "1.1"),
("1.0.4", "1.04"),
("0.96", "0.96"),
])
eggs = forms.CharField()
project = forms.CharField()
extra_paths = forms.CharField()
fcgi = forms.BooleanField()
wsgi = forms.BooleanField()
recipe_form_map = {
'djangorecipe': DjangoRecipeForm
}
|
61cf5a4ab4d7b9e0cb95925acc633aa7cb156d59
|
taggit/models.py
|
taggit/models.py
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField()
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField(unique=True, max_length=100)
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
Make slug unique and fix the max_length.
|
Make slug unique and fix the max_length.
|
Python
|
bsd-3-clause
|
izquierdo/django-taggit,theatlantic/django-taggit2,doselect/django-taggit,gem/django-taggit,theatlantic/django-taggit,Maplecroft/django-taggit,benjaminrigaud/django-taggit,kminkov/django-taggit,twig/django-taggit,cimani/django-taggit,decibyte/django-taggit,IRI-Research/django-taggit,eugena/django-taggit,theatlantic/django-taggit2,laanlabs/django-taggit,kaedroho/django-taggit,orbitvu/django-taggit,7kfpun/django-taggit,adrian-sgn/django-taggit,guoqiao/django-taggit,tamarmot/django-taggit,vhf/django-taggit,nealtodd/django-taggit,decibyte/django-taggit,theatlantic/django-taggit,Eksmo/django-taggit
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField()
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
Make slug unique and fix the max_length.
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField(unique=True, max_length=100)
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
<commit_before>from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField()
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
<commit_msg>Make slug unique and fix the max_length.<commit_after>
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField(unique=True, max_length=100)
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField()
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
Make slug unique and fix the max_length.from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField(unique=True, max_length=100)
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
<commit_before>from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField()
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
<commit_msg>Make slug unique and fix the max_length.<commit_after>from django.contrib.contenttypes.models import ContentType
from django.contrib.contenttypes.generic import GenericForeignKey
from django.db import models
from django.template.defaultfilters import slugify
class Tag(models.Model):
name = models.CharField(max_length=100)
slug = models.SlugField(unique=True, max_length=100)
def __unicode__(self):
return self.name
def save(self, *args, **kwargs):
if not self.pk:
self.slug = slugify(self.name)
super(Tag, self).save(*args, **kwargs)
class TaggedItem(models.Model):
object_id = models.IntegerField()
content_type = models.ForeignKey(ContentType)
content_object = GenericForeignKey()
tag = models.ForeignKey(Tag, related_name="items")
def __unicode__(self):
return "%s tagged with %s" % (self.content_object, self.tag)
|
9afc0f35b718e11418c22e2f60e07f7a9ee9aaa3
|
core/commands/log_graph.py
|
core/commands/log_graph.py
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
Add `--all` to graph command to show other branches (esp. origin)
|
Add `--all` to graph command to show other branches (esp. origin)
|
Python
|
mit
|
asfaltboy/GitSavvy,asfaltboy/GitSavvy,jmanuel1/GitSavvy,divmain/GitSavvy,ddevlin/GitSavvy,theiviaxx/GitSavvy,divmain/GitSavvy,stoivo/GitSavvy,dvcrn/GitSavvy,theiviaxx/GitSavvy,ralic/GitSavvy,ddevlin/GitSavvy,ralic/GitSavvy,divmain/GitSavvy,stoivo/GitSavvy,dreki/GitSavvy,stoivo/GitSavvy,asfaltboy/GitSavvy,ddevlin/GitSavvy,dvcrn/GitSavvy,jmanuel1/GitSavvy,dreki/GitSavvy
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
Add `--all` to graph command to show other branches (esp. origin)
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
<commit_before>from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
<commit_msg>Add `--all` to graph command to show other branches (esp. origin)<commit_after>
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
Add `--all` to graph command to show other branches (esp. origin)from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
<commit_before>from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
<commit_msg>Add `--all` to graph command to show other branches (esp. origin)<commit_after>from sublime_plugin import WindowCommand, TextCommand
from ..git_command import GitCommand
LOG_GRAPH_TITLE = "GRAPH"
class GsLogGraphCommand(WindowCommand, GitCommand):
"""
Open a new window displaying an ASCII-graphic representation
of the repo's branch relationships.
"""
def run(self):
repo_path = self.repo_path
view = self.window.new_file()
view.settings().set("git_savvy.log_graph_view", True)
view.settings().set("git_savvy.repo_path", repo_path)
view.set_name(LOG_GRAPH_TITLE)
view.set_scratch(True)
view.set_read_only(True)
view.run_command("gs_log_graph_initialize")
class GsLogGraphInitializeCommand(TextCommand, GitCommand):
def run(self, edit):
branch_graph = self.git("log", "--oneline", "--graph", "--all", "--decorate")
self.view.run_command("gs_replace_view_text", {"text": branch_graph})
|
7db62eefb0601098db6eff9becf566f51fee9be4
|
pythonforandroid/recipes/setuptools/__init__.py
|
pythonforandroid/recipes/setuptools/__init__.py
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
Use zip instead of tar.gz to obtain setuptools
|
Use zip instead of tar.gz to obtain setuptools
Newer releases of setuptools on pypi are only available as zip archives.
In order to support higher versions of setuptools, zip should be
preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0
will fail.
|
Python
|
mit
|
kivy/python-for-android,rnixx/python-for-android,germn/python-for-android,rnixx/python-for-android,PKRoma/python-for-android,rnixx/python-for-android,rnixx/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,germn/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,kronenpj/python-for-android,PKRoma/python-for-android,germn/python-for-android,rnixx/python-for-android,germn/python-for-android,kronenpj/python-for-android,kivy/python-for-android,germn/python-for-android,rnixx/python-for-android,kivy/python-for-android,kivy/python-for-android,PKRoma/python-for-android,germn/python-for-android,kivy/python-for-android,kronenpj/python-for-android
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
Use zip instead of tar.gz to obtain setuptools
Newer releases of setuptools on pypi are only available as zip archives.
In order to support higher versions of setuptools, zip should be
preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0
will fail.
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
<commit_before>from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
<commit_msg>Use zip instead of tar.gz to obtain setuptools
Newer releases of setuptools on pypi are only available as zip archives.
In order to support higher versions of setuptools, zip should be
preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0
will fail.<commit_after>
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
Use zip instead of tar.gz to obtain setuptools
Newer releases of setuptools on pypi are only available as zip archives.
In order to support higher versions of setuptools, zip should be
preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0
will fail.from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
<commit_before>from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.tar.gz'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
<commit_msg>Use zip instead of tar.gz to obtain setuptools
Newer releases of setuptools on pypi are only available as zip archives.
In order to support higher versions of setuptools, zip should be
preferred over tar.gz. Otherwise, installing e.g. setuptools==40.0.0
will fail.<commit_after>from pythonforandroid.recipe import PythonRecipe
class SetuptoolsRecipe(PythonRecipe):
version = '18.3.1'
url = 'https://pypi.python.org/packages/source/s/setuptools/setuptools-{version}.zip'
depends = [('python2', 'python3crystax')]
call_hostpython_via_targetpython = False
install_in_hostpython = True
recipe = SetuptoolsRecipe()
|
b3757884bdaa6e488d54ee51f943dbb3578ea469
|
stores/forms.py
|
stores/forms.py
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
max_length=4,
widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")})
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
widget=forms.TextInput(
attrs={'placeholder': _("Enter your postcode or suburb...")}
)
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
Remove limit for search field
|
Remove limit for search field
|
Python
|
bsd-3-clause
|
django-oscar/django-oscar-stores,django-oscar/django-oscar-stores,django-oscar/django-oscar-stores
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
max_length=4,
widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")})
)
state = forms.ChoiceField(choices=STATE_CHOICES)Remove limit for search field
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
widget=forms.TextInput(
attrs={'placeholder': _("Enter your postcode or suburb...")}
)
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
<commit_before>from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
max_length=4,
widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")})
)
state = forms.ChoiceField(choices=STATE_CHOICES)<commit_msg>Remove limit for search field<commit_after>
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
widget=forms.TextInput(
attrs={'placeholder': _("Enter your postcode or suburb...")}
)
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
max_length=4,
widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")})
)
state = forms.ChoiceField(choices=STATE_CHOICES)Remove limit for search fieldfrom django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
widget=forms.TextInput(
attrs={'placeholder': _("Enter your postcode or suburb...")}
)
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
<commit_before>from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
max_length=4,
widget=forms.TextInput(attrs={'placeholder': _("Enter your postcode or suburb...")})
)
state = forms.ChoiceField(choices=STATE_CHOICES)<commit_msg>Remove limit for search field<commit_after>from django import forms
from django.db.models import get_model
from django.utils.translation import ugettext as _
StoreAddress = get_model('stores', 'StoreAddress')
class StoreSearchForm(forms.Form):
STATE_CHOICES = (
(_('VIC'), _('Victoria')),
(_('NSW'), _('New South Wales')),
(_('SA'), _('South Australia')),
(_('TAS'), _('Tasmania')),
(_('QLD'), _('Queensland')),
(_('NT'), _('Northern Territory')),
)
location = forms.CharField(widget=forms.HiddenInput)
store_search = forms.CharField(
widget=forms.TextInput(
attrs={'placeholder': _("Enter your postcode or suburb...")}
)
)
state = forms.ChoiceField(choices=STATE_CHOICES)
|
fe50886a42bf7fa5e3217134e1f7a732960ab2d9
|
nbgrader/tests/apps/test_nbgrader_generate_config.py
|
nbgrader/tests/apps/test_nbgrader_generate_config.py
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
with open("nbgrader_config.py") as f:
contents = f.read()
# This was missing in issue #1089
assert "AssignLatePenalties" in contents
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
Add assertion for issue gh-1089
|
Add assertion for issue gh-1089
|
Python
|
bsd-3-clause
|
jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jupyter/nbgrader,jhamrick/nbgrader,jhamrick/nbgrader,jhamrick/nbgrader
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
Add assertion for issue gh-1089
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
with open("nbgrader_config.py") as f:
contents = f.read()
# This was missing in issue #1089
assert "AssignLatePenalties" in contents
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
<commit_before>import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
<commit_msg>Add assertion for issue gh-1089<commit_after>
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
with open("nbgrader_config.py") as f:
contents = f.read()
# This was missing in issue #1089
assert "AssignLatePenalties" in contents
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
Add assertion for issue gh-1089import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
with open("nbgrader_config.py") as f:
contents = f.read()
# This was missing in issue #1089
assert "AssignLatePenalties" in contents
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
<commit_before>import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
<commit_msg>Add assertion for issue gh-1089<commit_after>import os
from .. import run_nbgrader
from .base import BaseTestApp
class TestNbGraderGenerateConfig(BaseTestApp):
def test_help(self):
"""Does the help display without error?"""
run_nbgrader(["generate_config", "--help-all"])
def test_generate_config(self):
"""Is the config file properly generated?"""
# it already exists, because we create it in conftest.py
os.remove("nbgrader_config.py")
# try recreating it
run_nbgrader(["generate_config"])
assert os.path.isfile("nbgrader_config.py")
with open("nbgrader_config.py") as f:
contents = f.read()
# This was missing in issue #1089
assert "AssignLatePenalties" in contents
# does it fail if it already exists?
run_nbgrader(["generate_config"], retcode=1)
|
5d673ed6cf8f810ee22bc701f6927f9580b4ec34
|
rest-api/config.py
|
rest-api/config.py
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
STAGING_CLIENT_ID = '101582376895035372811'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
Add the staging client id too.
|
Add the staging client id too.
|
Python
|
bsd-3-clause
|
all-of-us/raw-data-repository,all-of-us/raw-data-repository,all-of-us/raw-data-repository
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
Add the staging client id too.
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
STAGING_CLIENT_ID = '101582376895035372811'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
<commit_before>"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
<commit_msg>Add the staging client id too.<commit_after>
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
STAGING_CLIENT_ID = '101582376895035372811'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
Add the staging client id too."""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
STAGING_CLIENT_ID = '101582376895035372811'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
<commit_before>"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
<commit_msg>Add the staging client id too.<commit_after>"""Configuration parameters.
Contains things such as the database to connect to.
"""
CLOUDSQL_INSTANCE = 'pmi-drc-api-test:us-central1:pmi-rdr'
CLOUDSQL_SOCKET = '/cloudsql/' + CLOUDSQL_INSTANCE
CLOUDSQL_USER = 'api'
PYTHON_TEST_CLIENT_ID = '116540421226121250670'
STAGING_CLIENT_ID = '101582376895035372811'
ALLOWED_CLIENT_IDS = [PYTHON_TEST_CLIENT_ID, STAGING_CLIENT_ID]
# TODO: Move all authentication into the datastore.
ALLOWED_USERS = [
'test-client@pmi-rdr-api-test.iam.gserviceaccount.com',
'pmi-hpo-staging@appspot.gserviceaccount.com',
]
|
d8a861c47df6b41c27f2ec43474766284ba728af
|
bot/logger/message_sender/reusable/limiter/group.py
|
bot/logger/message_sender/reusable/limiter/group.py
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
def notify_about_to_send_message(self):
for limiter in self.limiters:
limiter.notify_about_to_send_message()
|
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
|
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
|
Python
|
agpl-3.0
|
alvarogzp/telegram-bot,alvarogzp/telegram-bot
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
def notify_about_to_send_message(self):
for limiter in self.limiters:
limiter.notify_about_to_send_message()
|
<commit_before>from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
<commit_msg>Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters<commit_after>
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
def notify_about_to_send_message(self):
for limiter in self.limiters:
limiter.notify_about_to_send_message()
|
from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limitersfrom bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
def notify_about_to_send_message(self):
for limiter in self.limiters:
limiter.notify_about_to_send_message()
|
<commit_before>from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
<commit_msg>Make ReusableMessageGroup broadcast the notify_about_to_send_message to all limiters<commit_after>from bot.logger.message_sender.message_builder import MessageBuilder
from bot.logger.message_sender.reusable.limiter import ReusableMessageLimiter
class ReusableMessageLimiterGroup(ReusableMessageLimiter):
def __init__(self, *limiters: ReusableMessageLimiter):
self.limiters = limiters
def should_issue_new_message_pre_add(self, new_text):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_pre_add(new_text))
def should_issue_new_message_post_add(self, builder: MessageBuilder):
return self.__any_limiter(lambda limiter: limiter.should_issue_new_message_post_add(builder))
def __any_limiter(self, func: callable):
return any((func(limiter) for limiter in self.limiters))
def notify_new_message_issued(self):
for limiter in self.limiters:
limiter.notify_new_message_issued()
def notify_about_to_send_message(self):
for limiter in self.limiters:
limiter.notify_about_to_send_message()
|
769fa2c0c777ac88702e6b3802de4909c8f8df22
|
sh_app/forms.py
|
sh_app/forms.py
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state', 'is_private')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
Remove isPrivate field on league form
|
Remove isPrivate field on league form
|
Python
|
mit
|
skill-huddle/skill-huddle,skill-huddle/skill-huddle
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state', 'is_private')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
Remove isPrivate field on league form
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
<commit_before>from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state', 'is_private')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
<commit_msg>Remove isPrivate field on league form<commit_after>
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state', 'is_private')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
Remove isPrivate field on league formfrom django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
<commit_before>from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state', 'is_private')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
<commit_msg>Remove isPrivate field on league form<commit_after>from django import forms
from django.forms import Textarea
from sh_app.models import User, SH_User, League, Suggestion
class UserForm(forms.ModelForm):
password = forms.CharField(widget=forms.PasswordInput())
class Meta:
model = User
fields = ('username', 'email', 'password')
class SH_UserForm(forms.ModelForm):
class Meta:
model = SH_User
fields = ('first_name', 'last_name')
class LeagueForm(forms.ModelForm):
class Meta:
model = League
fields = ('name', 'description', 'country', 'city', 'state')
help_texts = {
'description': 'Maximum {} characters'.format(League._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea
}
class SuggestionForm(forms.ModelForm):
class Meta:
model = Suggestion
fields = ('name', 'description', 'voting_ends')
help_texts = {
'description': 'Maximum {} characters'.format(Suggestion._meta.get_field('description').max_length)
}
widgets = {
'description': Textarea,
}
|
6dd52ba31141f28e1f37e32f8c3de6932ed49b4f
|
make_mozilla/base/tests/assertions.py
|
make_mozilla/base/tests/assertions.py
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_(resolved_route.func is view_function)
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function_or_class, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class))
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
Allow assert_routing to cope with Feed instances used as route endpoints
|
Allow assert_routing to cope with Feed instances used as route endpoints
|
Python
|
bsd-3-clause
|
mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org,mozilla/make.mozilla.org
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_(resolved_route.func is view_function)
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
Allow assert_routing to cope with Feed instances used as route endpoints
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function_or_class, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class))
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
<commit_before>from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_(resolved_route.func is view_function)
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
<commit_msg>Allow assert_routing to cope with Feed instances used as route endpoints<commit_after>
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function_or_class, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class))
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_(resolved_route.func is view_function)
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
Allow assert_routing to cope with Feed instances used as route endpointsfrom nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function_or_class, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class))
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
<commit_before>from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_(resolved_route.func is view_function)
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
<commit_msg>Allow assert_routing to cope with Feed instances used as route endpoints<commit_after>from nose.tools import eq_, ok_
from django.core.urlresolvers import resolve, reverse
def assert_routing(url, view_function_or_class, name = '', kwargs = {}):
resolved_route = resolve(url)
ok_((resolved_route.func is view_function_or_class) or (type(resolved_route.func) is view_function_or_class))
if kwargs:
eq_(resolved_route.kwargs, kwargs)
if name:
eq_(reverse(name, kwargs = kwargs), url)
def assert_redirects_to_named_url(response, name, kwargs = {}, permanent = False):
status_codes = {True: 301, False: 302}
expected_redirect_url = reverse(name, kwargs = kwargs)
eq_(response.status_code, status_codes[permanent])
eq_(response['Location'], expected_redirect_url)
|
8f530fac6e16dcae18aae8696b5a30117137f9f3
|
rest/messages/sms-conversation-tracking/example-1.py
|
rest/messages/sms-conversation-tracking/example-1.py
|
from flask import Flask, request, redirect, session
import twilio.twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
counter = session.get('counter', 0)
# increment the counter
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
message = "".join([name, " has messaged ", request.values.get('To'), " ",
str(counter), " times."])
resp = twilio.twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
from flask import Flask, request, session
from twilio import twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
# Increment the counter
counter = session.get('counter', 0)
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
# Build our reply
message = '{} has messaged {} {} times.' \
.format(name, request.values.get('To'), counter)
# Put it in a TwiML response
resp = twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
Clean up Python example for sms-conversation-tracking
|
Clean up Python example for sms-conversation-tracking
|
Python
|
mit
|
teoreteetik/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,teoreteetik/api-snippets,TwilioDevEd/api-snippets,TwilioDevEd/api-snippets
|
from flask import Flask, request, redirect, session
import twilio.twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
counter = session.get('counter', 0)
# increment the counter
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
message = "".join([name, " has messaged ", request.values.get('To'), " ",
str(counter), " times."])
resp = twilio.twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)Clean up Python example for sms-conversation-tracking
|
from flask import Flask, request, session
from twilio import twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
# Increment the counter
counter = session.get('counter', 0)
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
# Build our reply
message = '{} has messaged {} {} times.' \
.format(name, request.values.get('To'), counter)
# Put it in a TwiML response
resp = twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
<commit_before>from flask import Flask, request, redirect, session
import twilio.twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
counter = session.get('counter', 0)
# increment the counter
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
message = "".join([name, " has messaged ", request.values.get('To'), " ",
str(counter), " times."])
resp = twilio.twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)<commit_msg>Clean up Python example for sms-conversation-tracking<commit_after>
|
from flask import Flask, request, session
from twilio import twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
# Increment the counter
counter = session.get('counter', 0)
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
# Build our reply
message = '{} has messaged {} {} times.' \
.format(name, request.values.get('To'), counter)
# Put it in a TwiML response
resp = twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
from flask import Flask, request, redirect, session
import twilio.twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
counter = session.get('counter', 0)
# increment the counter
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
message = "".join([name, " has messaged ", request.values.get('To'), " ",
str(counter), " times."])
resp = twilio.twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)Clean up Python example for sms-conversation-trackingfrom flask import Flask, request, session
from twilio import twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
# Increment the counter
counter = session.get('counter', 0)
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
# Build our reply
message = '{} has messaged {} {} times.' \
.format(name, request.values.get('To'), counter)
# Put it in a TwiML response
resp = twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
<commit_before>from flask import Flask, request, redirect, session
import twilio.twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
counter = session.get('counter', 0)
# increment the counter
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
message = "".join([name, " has messaged ", request.values.get('To'), " ",
str(counter), " times."])
resp = twilio.twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)<commit_msg>Clean up Python example for sms-conversation-tracking<commit_after>from flask import Flask, request, session
from twilio import twiml
# The session object makes use of a secret key.
SECRET_KEY = 'a secret key'
app = Flask(__name__)
app.config.from_object(__name__)
# Try adding your own number to this list!
callers = {
"+14158675309": "Rey",
"+14158675310": "Finn",
"+14158675311": "Chewy",
}
@app.route("/", methods=['GET', 'POST'])
def hello():
"""Respond with the number of text messages sent between two parties."""
# Increment the counter
counter = session.get('counter', 0)
counter += 1
# Save the new counter value in the session
session['counter'] = counter
from_number = request.values.get('From')
if from_number in callers:
name = callers[from_number]
else:
name = "Friend"
# Build our reply
message = '{} has messaged {} {} times.' \
.format(name, request.values.get('To'), counter)
# Put it in a TwiML response
resp = twiml.Response()
resp.message(message)
return str(resp)
if __name__ == "__main__":
app.run(debug=True)
|
e6873d3d40d868e743a239c6a74a345c2999541e
|
dusty_coffin/elasticsearch_custom/edit_spreadsheet.py
|
dusty_coffin/elasticsearch_custom/edit_spreadsheet.py
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
print(d)
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
Edit spreadsheet is now fixed.
|
Edit spreadsheet is now fixed.
|
Python
|
mit
|
bhillmann/dusty_coffin,bhillmann/dusty_coffin,bhillmann/dusty_coffin,bhillmann/dusty_coffin
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
Edit spreadsheet is now fixed.
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
print(d)
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
<commit_before>import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
<commit_msg>Edit spreadsheet is now fixed.<commit_after>
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
print(d)
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
Edit spreadsheet is now fixed.import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
print(d)
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
<commit_before>import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=0Ahf71UaPpMOSdGl0NnQtSFgyVFpvSmV3R2JobzVmZHc&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
<commit_msg>Edit spreadsheet is now fixed.<commit_after>import urllib.request
from io import BytesIO
import pandas as pd
import ujson
from elasticsearch import Elasticsearch
csv = urllib.request.urlopen("https://docs.google.com/spreadsheet/pub?key=1h1udf_H073YaVlZs0fkYUf9dC6KbEZAhF1veeLExyXo&gid=937170620&output=csv").read()
bio = BytesIO(csv)
csv_pd = pd.DataFrame.from_csv(bio)
json_objs = csv_pd.reset_index().to_json(orient='records')
dict_array= ujson.loads(json_objs)
# Edit to point to elasticsearch instance
es = Elasticsearch('ec2-52-10-17-100.us-west-2.compute.amazonaws.com:9200')
# ignore 404 and 400
es.indices.delete(index='datasets_index', ignore=[400, 404])
for d in dict_array:
print(d)
res = es.index(index="datasets_index", doc_type="dataset", body=d)
print(res['created'])
|
8eb0b7fcd6ffb81d6b0fc69cb31c7625550583d7
|
targetrupypy.py
|
targetrupypy.py
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "rupypy-c"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "./bin/topaz"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
Move towards a normal bin directory.
|
Move towards a normal bin directory.
|
Python
|
bsd-3-clause
|
babelsberg/babelsberg-r,topazproject/topaz,babelsberg/babelsberg-r,kachick/topaz,kachick/topaz,babelsberg/babelsberg-r,babelsberg/babelsberg-r,babelsberg/babelsberg-r,kachick/topaz,topazproject/topaz,topazproject/topaz,topazproject/topaz
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "rupypy-c"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()Move towards a normal bin directory.
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "./bin/topaz"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
<commit_before>from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "rupypy-c"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()<commit_msg>Move towards a normal bin directory.<commit_after>
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "./bin/topaz"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "rupypy-c"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()Move towards a normal bin directory.from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "./bin/topaz"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
<commit_before>from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "rupypy-c"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()<commit_msg>Move towards a normal bin directory.<commit_after>from pypy.jit.codewriter.policy import JitPolicy
from rupypy.main import entry_point
def target(driver, args):
driver.exe_name = "./bin/topaz"
return entry_point, None
def jitpolicy(driver):
return JitPolicy()
|
39561a89ea497776d980d3eda97fc2f75493528f
|
internal_social_auth/views.py
|
internal_social_auth/views.py
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
Clean up the AuthComplete API a little
|
Clean up the AuthComplete API a little
|
Python
|
bsd-2-clause
|
incuna/incuna-internal-social-auth
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
Clean up the AuthComplete API a little
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
<commit_before>import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
<commit_msg>Clean up the AuthComplete API a little<commit_after>
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
Clean up the AuthComplete API a littleimport logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
<commit_before>import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, "Your Google Apps domain isn't authorized for this app")
return HttpResponseRedirect('/')
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
<commit_msg>Clean up the AuthComplete API a little<commit_after>import logging
from django.contrib import messages
from django.http import HttpResponseRedirect, HttpResponse
from django.utils.encoding import force_text
from django.views.generic.base import View
from social_auth.exceptions import AuthFailed
from social_auth.views import complete
logger = logging.getLogger(__name__)
class AuthComplete(View):
def get(self, request, *args, **kwargs):
backend = kwargs.pop('backend')
try:
return complete(request, backend, *args, **kwargs)
except AuthFailed as e:
logger.error(e)
messages.error(request, self.get_error_message())
return HttpResponseRedirect(self.get_faiure_url())
def get_error_message(self):
if self.error_message:
return self.error_message
return "Your Google Apps domain isn't authorized for this app"
def get_failure_url(self):
if self.failure_url:
return force_text(self.failure_url)
return '/'
class LoginError(View):
def get(self, request, *args, **kwargs):
return HttpResponse(status=401)
|
97b000547898d6aa3006f07cf2ef9d8656a67865
|
conllu/__init__.py
|
conllu/__init__.py
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def _iter_sents(in_file):
buf = []
for line in in_file:
if line == "\n":
yield "".join(buf)[:-1]
buf = []
else:
buf.append(line)
if buf:
yield "".join(buf)
def parse_incr(in_file, fields=None):
for sentence in _iter_sents(in_file):
yield TokenList(*parse_token_and_metadata(sentence, fields=fields))
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
def parse_tree_incr(in_file):
for tokenlist in parse_incr(in_file):
yield tokenlist.to_tree()
|
Add incremental parsing versions of parse and parse_tree
|
Add incremental parsing versions of parse and parse_tree
|
Python
|
mit
|
EmilStenstrom/conllu
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
Add incremental parsing versions of parse and parse_tree
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def _iter_sents(in_file):
buf = []
for line in in_file:
if line == "\n":
yield "".join(buf)[:-1]
buf = []
else:
buf.append(line)
if buf:
yield "".join(buf)
def parse_incr(in_file, fields=None):
for sentence in _iter_sents(in_file):
yield TokenList(*parse_token_and_metadata(sentence, fields=fields))
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
def parse_tree_incr(in_file):
for tokenlist in parse_incr(in_file):
yield tokenlist.to_tree()
|
<commit_before>from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
<commit_msg>Add incremental parsing versions of parse and parse_tree<commit_after>
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def _iter_sents(in_file):
buf = []
for line in in_file:
if line == "\n":
yield "".join(buf)[:-1]
buf = []
else:
buf.append(line)
if buf:
yield "".join(buf)
def parse_incr(in_file, fields=None):
for sentence in _iter_sents(in_file):
yield TokenList(*parse_token_and_metadata(sentence, fields=fields))
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
def parse_tree_incr(in_file):
for tokenlist in parse_incr(in_file):
yield tokenlist.to_tree()
|
from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
Add incremental parsing versions of parse and parse_treefrom __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def _iter_sents(in_file):
buf = []
for line in in_file:
if line == "\n":
yield "".join(buf)[:-1]
buf = []
else:
buf.append(line)
if buf:
yield "".join(buf)
def parse_incr(in_file, fields=None):
for sentence in _iter_sents(in_file):
yield TokenList(*parse_token_and_metadata(sentence, fields=fields))
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
def parse_tree_incr(in_file):
for tokenlist in parse_incr(in_file):
yield tokenlist.to_tree()
|
<commit_before>from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
<commit_msg>Add incremental parsing versions of parse and parse_tree<commit_after>from __future__ import unicode_literals
from conllu.models import TokenList
from conllu.parser import parse_token_and_metadata
def parse(data, fields=None):
return [
TokenList(*parse_token_and_metadata(sentence, fields=fields))
for sentence in data.split("\n\n")
if sentence
]
def _iter_sents(in_file):
buf = []
for line in in_file:
if line == "\n":
yield "".join(buf)[:-1]
buf = []
else:
buf.append(line)
if buf:
yield "".join(buf)
def parse_incr(in_file, fields=None):
for sentence in _iter_sents(in_file):
yield TokenList(*parse_token_and_metadata(sentence, fields=fields))
def parse_tree(data):
tokenlists = parse(data)
sentences = []
for tokenlist in tokenlists:
sentences.append(tokenlist.to_tree())
return sentences
def parse_tree_incr(in_file):
for tokenlist in parse_incr(in_file):
yield tokenlist.to_tree()
|
9e3a6190b2dcfd7de03ef5c974b400a51219839e
|
pyof/v0x04/symmetric/hello.py
|
pyof/v0x04/symmetric/hello.py
|
"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.v0x01.symmetric.hello import Hello
__all__ = ('Hello',)
|
"""Defines Hello message."""
# System imports
from enum import Enum
from pyof.foundation.base import GenericMessage, GenericStruct
from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16
from pyof.v0x04.common.header import Header, Type
# Third-party imports
__all__ = ('Hello', 'HelloElemHeader', 'HelloElemType',
'HelloElemVersionbitmap', 'ListOfHelloElements')
# Enums
class HelloElemType(Enum):
"""Hello element types."""
#: Bitmap of version supported.
OFPHET_VERSIONBITMAP = 1
# Classes
class HelloElemHeader(GenericStruct):
"""Common header for all Hello Elements."""
element_type = UBInt16(enum_ref=HelloElemType)
length = UBInt16()
def __init__(self, element_type=None, length=None):
"""The constructor just assigns parameters to object attributes.
Args:
element_type: One of OFPHET_*.
length: Length in bytes of the element, including this header,
excluding padding.
"""
super().__init__()
self.element_type = element_type
self.length = length
class ListOfHelloElements(FixedTypeList):
"""List of Hello elements.
Represented by instances of HelloElemHeader and used on Hello
objects.
"""
def __init__(self, items=None):
"""The constructor just assigns parameters to object attributes.
Args:
items (HelloElemHeader): Instance or a list of instances.
"""
super().__init__(pyof_class=HelloElemHeader, items=items)
class Hello(GenericMessage):
"""OpenFlow Hello Message OFPT_HELLO.
This message includes zero or more hello elements having variable size.
Unknown element types must be ignored/skipped, to allow for future
extensions.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
#: Hello element list
elements = ListOfHelloElements()
def __init__(self, xid=None, elements=None):
"""The constructor takes the parameters below.
Args:
xid (int): xid to be used on the message header.
elements: List of elements - 0 or more
"""
super().__init__(xid)
self.elements = elements
class HelloElemVersionbitmap(HelloElemHeader):
"""Version bitmap Hello Element."""
#: List of bitmaps - supported versions
bitmaps = BinaryData()
def __init__(self, bitmaps=b''):
"""The constructor just assigns parameters to object attributes.
Args:
bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then
-Exactly (length + 7)/8*8 - (length) (between 0 and 7)
bytes of all-zero bytes.
"""
super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP,
length=None)
self.bitmaps = bitmaps
|
Add Hello class and related classes for v0x04
|
Add Hello class and related classes for v0x04
Fix #302
Fix #303
|
Python
|
mit
|
cemsbr/python-openflow,kytos/python-openflow
|
"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.v0x01.symmetric.hello import Hello
__all__ = ('Hello',)
Add Hello class and related classes for v0x04
Fix #302
Fix #303
|
"""Defines Hello message."""
# System imports
from enum import Enum
from pyof.foundation.base import GenericMessage, GenericStruct
from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16
from pyof.v0x04.common.header import Header, Type
# Third-party imports
__all__ = ('Hello', 'HelloElemHeader', 'HelloElemType',
'HelloElemVersionbitmap', 'ListOfHelloElements')
# Enums
class HelloElemType(Enum):
"""Hello element types."""
#: Bitmap of version supported.
OFPHET_VERSIONBITMAP = 1
# Classes
class HelloElemHeader(GenericStruct):
"""Common header for all Hello Elements."""
element_type = UBInt16(enum_ref=HelloElemType)
length = UBInt16()
def __init__(self, element_type=None, length=None):
"""The constructor just assigns parameters to object attributes.
Args:
element_type: One of OFPHET_*.
length: Length in bytes of the element, including this header,
excluding padding.
"""
super().__init__()
self.element_type = element_type
self.length = length
class ListOfHelloElements(FixedTypeList):
"""List of Hello elements.
Represented by instances of HelloElemHeader and used on Hello
objects.
"""
def __init__(self, items=None):
"""The constructor just assigns parameters to object attributes.
Args:
items (HelloElemHeader): Instance or a list of instances.
"""
super().__init__(pyof_class=HelloElemHeader, items=items)
class Hello(GenericMessage):
"""OpenFlow Hello Message OFPT_HELLO.
This message includes zero or more hello elements having variable size.
Unknown element types must be ignored/skipped, to allow for future
extensions.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
#: Hello element list
elements = ListOfHelloElements()
def __init__(self, xid=None, elements=None):
"""The constructor takes the parameters below.
Args:
xid (int): xid to be used on the message header.
elements: List of elements - 0 or more
"""
super().__init__(xid)
self.elements = elements
class HelloElemVersionbitmap(HelloElemHeader):
"""Version bitmap Hello Element."""
#: List of bitmaps - supported versions
bitmaps = BinaryData()
def __init__(self, bitmaps=b''):
"""The constructor just assigns parameters to object attributes.
Args:
bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then
-Exactly (length + 7)/8*8 - (length) (between 0 and 7)
bytes of all-zero bytes.
"""
super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP,
length=None)
self.bitmaps = bitmaps
|
<commit_before>"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.v0x01.symmetric.hello import Hello
__all__ = ('Hello',)
<commit_msg>Add Hello class and related classes for v0x04
Fix #302
Fix #303<commit_after>
|
"""Defines Hello message."""
# System imports
from enum import Enum
from pyof.foundation.base import GenericMessage, GenericStruct
from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16
from pyof.v0x04.common.header import Header, Type
# Third-party imports
__all__ = ('Hello', 'HelloElemHeader', 'HelloElemType',
'HelloElemVersionbitmap', 'ListOfHelloElements')
# Enums
class HelloElemType(Enum):
"""Hello element types."""
#: Bitmap of version supported.
OFPHET_VERSIONBITMAP = 1
# Classes
class HelloElemHeader(GenericStruct):
"""Common header for all Hello Elements."""
element_type = UBInt16(enum_ref=HelloElemType)
length = UBInt16()
def __init__(self, element_type=None, length=None):
"""The constructor just assigns parameters to object attributes.
Args:
element_type: One of OFPHET_*.
length: Length in bytes of the element, including this header,
excluding padding.
"""
super().__init__()
self.element_type = element_type
self.length = length
class ListOfHelloElements(FixedTypeList):
"""List of Hello elements.
Represented by instances of HelloElemHeader and used on Hello
objects.
"""
def __init__(self, items=None):
"""The constructor just assigns parameters to object attributes.
Args:
items (HelloElemHeader): Instance or a list of instances.
"""
super().__init__(pyof_class=HelloElemHeader, items=items)
class Hello(GenericMessage):
"""OpenFlow Hello Message OFPT_HELLO.
This message includes zero or more hello elements having variable size.
Unknown element types must be ignored/skipped, to allow for future
extensions.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
#: Hello element list
elements = ListOfHelloElements()
def __init__(self, xid=None, elements=None):
"""The constructor takes the parameters below.
Args:
xid (int): xid to be used on the message header.
elements: List of elements - 0 or more
"""
super().__init__(xid)
self.elements = elements
class HelloElemVersionbitmap(HelloElemHeader):
"""Version bitmap Hello Element."""
#: List of bitmaps - supported versions
bitmaps = BinaryData()
def __init__(self, bitmaps=b''):
"""The constructor just assigns parameters to object attributes.
Args:
bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then
-Exactly (length + 7)/8*8 - (length) (between 0 and 7)
bytes of all-zero bytes.
"""
super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP,
length=None)
self.bitmaps = bitmaps
|
"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.v0x01.symmetric.hello import Hello
__all__ = ('Hello',)
Add Hello class and related classes for v0x04
Fix #302
Fix #303"""Defines Hello message."""
# System imports
from enum import Enum
from pyof.foundation.base import GenericMessage, GenericStruct
from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16
from pyof.v0x04.common.header import Header, Type
# Third-party imports
__all__ = ('Hello', 'HelloElemHeader', 'HelloElemType',
'HelloElemVersionbitmap', 'ListOfHelloElements')
# Enums
class HelloElemType(Enum):
"""Hello element types."""
#: Bitmap of version supported.
OFPHET_VERSIONBITMAP = 1
# Classes
class HelloElemHeader(GenericStruct):
"""Common header for all Hello Elements."""
element_type = UBInt16(enum_ref=HelloElemType)
length = UBInt16()
def __init__(self, element_type=None, length=None):
"""The constructor just assigns parameters to object attributes.
Args:
element_type: One of OFPHET_*.
length: Length in bytes of the element, including this header,
excluding padding.
"""
super().__init__()
self.element_type = element_type
self.length = length
class ListOfHelloElements(FixedTypeList):
"""List of Hello elements.
Represented by instances of HelloElemHeader and used on Hello
objects.
"""
def __init__(self, items=None):
"""The constructor just assigns parameters to object attributes.
Args:
items (HelloElemHeader): Instance or a list of instances.
"""
super().__init__(pyof_class=HelloElemHeader, items=items)
class Hello(GenericMessage):
"""OpenFlow Hello Message OFPT_HELLO.
This message includes zero or more hello elements having variable size.
Unknown element types must be ignored/skipped, to allow for future
extensions.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
#: Hello element list
elements = ListOfHelloElements()
def __init__(self, xid=None, elements=None):
"""The constructor takes the parameters below.
Args:
xid (int): xid to be used on the message header.
elements: List of elements - 0 or more
"""
super().__init__(xid)
self.elements = elements
class HelloElemVersionbitmap(HelloElemHeader):
"""Version bitmap Hello Element."""
#: List of bitmaps - supported versions
bitmaps = BinaryData()
def __init__(self, bitmaps=b''):
"""The constructor just assigns parameters to object attributes.
Args:
bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then
-Exactly (length + 7)/8*8 - (length) (between 0 and 7)
bytes of all-zero bytes.
"""
super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP,
length=None)
self.bitmaps = bitmaps
|
<commit_before>"""Defines Hello message."""
# System imports
# Third-party imports
from pyof.v0x01.symmetric.hello import Hello
__all__ = ('Hello',)
<commit_msg>Add Hello class and related classes for v0x04
Fix #302
Fix #303<commit_after>"""Defines Hello message."""
# System imports
from enum import Enum
from pyof.foundation.base import GenericMessage, GenericStruct
from pyof.foundation.basic_types import BinaryData, FixedTypeList, UBInt16
from pyof.v0x04.common.header import Header, Type
# Third-party imports
__all__ = ('Hello', 'HelloElemHeader', 'HelloElemType',
'HelloElemVersionbitmap', 'ListOfHelloElements')
# Enums
class HelloElemType(Enum):
"""Hello element types."""
#: Bitmap of version supported.
OFPHET_VERSIONBITMAP = 1
# Classes
class HelloElemHeader(GenericStruct):
"""Common header for all Hello Elements."""
element_type = UBInt16(enum_ref=HelloElemType)
length = UBInt16()
def __init__(self, element_type=None, length=None):
"""The constructor just assigns parameters to object attributes.
Args:
element_type: One of OFPHET_*.
length: Length in bytes of the element, including this header,
excluding padding.
"""
super().__init__()
self.element_type = element_type
self.length = length
class ListOfHelloElements(FixedTypeList):
"""List of Hello elements.
Represented by instances of HelloElemHeader and used on Hello
objects.
"""
def __init__(self, items=None):
"""The constructor just assigns parameters to object attributes.
Args:
items (HelloElemHeader): Instance or a list of instances.
"""
super().__init__(pyof_class=HelloElemHeader, items=items)
class Hello(GenericMessage):
"""OpenFlow Hello Message OFPT_HELLO.
This message includes zero or more hello elements having variable size.
Unknown element types must be ignored/skipped, to allow for future
extensions.
"""
header = Header(message_type=Type.OFPT_HELLO, length=8)
#: Hello element list
elements = ListOfHelloElements()
def __init__(self, xid=None, elements=None):
"""The constructor takes the parameters below.
Args:
xid (int): xid to be used on the message header.
elements: List of elements - 0 or more
"""
super().__init__(xid)
self.elements = elements
class HelloElemVersionbitmap(HelloElemHeader):
"""Version bitmap Hello Element."""
#: List of bitmaps - supported versions
bitmaps = BinaryData()
def __init__(self, bitmaps=b''):
"""The constructor just assigns parameters to object attributes.
Args:
bitmaps: -Exactly (length - 4) bytes containing the bitmaps,then
-Exactly (length + 7)/8*8 - (length) (between 0 and 7)
bytes of all-zero bytes.
"""
super().__init__(element_type=HelloElemType.OFPHET_VERSIONBITMAP,
length=None)
self.bitmaps = bitmaps
|
66ae5304c81d74e8f30e9274c90d0f83766744d7
|
datamodel/nodes/printer.py
|
datamodel/nodes/printer.py
|
import sys
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
sys.stdout.write(self._context)
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
import sys
import os
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
try:
sys.stdout.write(str(self._context))
sys.stdout.write(os.linesep)
except:
pass
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
Make sure we only write chars to stdout
|
Make sure we only write chars to stdout
|
Python
|
apache-2.0
|
csparpa/robograph,csparpa/robograph
|
import sys
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
sys.stdout.write(self._context)
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._contextMake sure we only write chars to stdout
|
import sys
import os
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
try:
sys.stdout.write(str(self._context))
sys.stdout.write(os.linesep)
except:
pass
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
<commit_before>import sys
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
sys.stdout.write(self._context)
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context<commit_msg>Make sure we only write chars to stdout<commit_after>
|
import sys
import os
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
try:
sys.stdout.write(str(self._context))
sys.stdout.write(os.linesep)
except:
pass
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
import sys
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
sys.stdout.write(self._context)
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._contextMake sure we only write chars to stdoutimport sys
import os
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
try:
sys.stdout.write(str(self._context))
sys.stdout.write(os.linesep)
except:
pass
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
<commit_before>import sys
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
sys.stdout.write(self._context)
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context<commit_msg>Make sure we only write chars to stdout<commit_after>import sys
import os
from datamodel.base import node
class ConsolePrinter(node.Node):
"""
This node prints on stdout its context and then returns it as output.
"""
def input(self, context):
self._context = context
def output(self):
try:
sys.stdout.write(str(self._context))
sys.stdout.write(os.linesep)
except:
pass
return self._context
def reset(self):
del self._context
class LogPrinter(node.Node):
"""
This node prints its context on a statically defined logger and then
returns it as output
"""
def __init__(self, logger, loglevel, stringify=False, name=None):
"""
:param logger: any logging.Logger subtype
:param loglevel: the log level
:param stringify: try to cast to str the context before passing it to
the logger
:param name: name of this node
"""
node.Node.__init__(self, name=name)
self._logger = logger
self._loglevel = loglevel
self._stringify = stringify
def input(self, context):
self._context = context
def output(self):
str_context = self._context
if self._stringify:
try:
str_context = str(self._context)
except:
pass # oops...
self._logger.log(self._loglevel, str_context)
return self._context
def reset(self):
del self._context
|
f7153fd88f07f99181f790a93559efd585272f18
|
nuxeo-drive-client/tests/test_copy.py
|
nuxeo-drive-client/tests/test_copy.py
|
from tests.common_unit_test import UnitTestCase
class TestCopy(UnitTestCase):
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
from tests.common_unit_test import RandomBug, UnitTestCase
class TestCopy(UnitTestCase):
@RandomBug('NXDRIVE-808', target='linux', repeat=5)
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
Add RandomBug for Linux on test_synchronize_remote_copy
|
NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy
|
Python
|
lgpl-2.1
|
ssdi-drive/nuxeo-drive,ssdi-drive/nuxeo-drive,ssdi-drive/nuxeo-drive
|
from tests.common_unit_test import UnitTestCase
class TestCopy(UnitTestCase):
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy
|
from tests.common_unit_test import RandomBug, UnitTestCase
class TestCopy(UnitTestCase):
@RandomBug('NXDRIVE-808', target='linux', repeat=5)
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
<commit_before>from tests.common_unit_test import UnitTestCase
class TestCopy(UnitTestCase):
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
<commit_msg>NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy<commit_after>
|
from tests.common_unit_test import RandomBug, UnitTestCase
class TestCopy(UnitTestCase):
@RandomBug('NXDRIVE-808', target='linux', repeat=5)
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
from tests.common_unit_test import UnitTestCase
class TestCopy(UnitTestCase):
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copyfrom tests.common_unit_test import RandomBug, UnitTestCase
class TestCopy(UnitTestCase):
@RandomBug('NXDRIVE-808', target='linux', repeat=5)
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
<commit_before>from tests.common_unit_test import UnitTestCase
class TestCopy(UnitTestCase):
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
<commit_msg>NXDRIVE-808: Add RandomBug for Linux on test_synchronize_remote_copy<commit_after>from tests.common_unit_test import RandomBug, UnitTestCase
class TestCopy(UnitTestCase):
@RandomBug('NXDRIVE-808', target='linux', repeat=5)
def test_synchronize_remote_copy(self):
local = self.local_client_1
remote = self.remote_document_client_1
# Create a file and a folder in the remote root workspace
remote.make_file('/', 'test.odt', 'Some content.')
remote.make_folder('/', 'Test folder')
# Launch ndrive and check synchronization
self.engine_1.start()
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/'))
self.assertTrue(local.exists('/Test folder'))
self.assertTrue(local.exists('/test.odt'))
# Copy the file to the folder remotely
remote.copy('/test.odt', '/Test folder')
# Launch ndrive and check synchronization
self.wait_sync(wait_for_async=True)
self.assertTrue(local.exists('/test.odt'))
self.assertEqual(local.get_content('/test.odt'),
'Some content.')
self.assertTrue(local.exists('/Test folder/test.odt'))
self.assertEqual(local.get_content('/Test folder/test.odt'),
'Some content.')
|
0dc72761a3b4b17098633df27fdbb70058afe311
|
geotrek/signage/migrations/0013_auto_20200423_1255.py
|
geotrek/signage/migrations/0013_auto_20200423_1255.py
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
]
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
def delete_force(apps, schema_editor):
# We can't import Infrastructure models directly as it may be a newer
# version than this migration expects. We use the historical version.
Blade = apps.get_model('signage', 'Blade')
for blade in Blade.objects.filter(deleted=True):
blade.delete()
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
migrations.RunPython(delete_force),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
]
|
Change order migration, user runpython instead
|
Change order migration, user runpython instead
|
Python
|
bsd-2-clause
|
makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin,GeotrekCE/Geotrek-admin,makinacorpus/Geotrek,makinacorpus/Geotrek,GeotrekCE/Geotrek-admin
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
]
Change order migration, user runpython instead
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
def delete_force(apps, schema_editor):
# We can't import Infrastructure models directly as it may be a newer
# version than this migration expects. We use the historical version.
Blade = apps.get_model('signage', 'Blade')
for blade in Blade.objects.filter(deleted=True):
blade.delete()
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
migrations.RunPython(delete_force),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
]
|
<commit_before># Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
]
<commit_msg>Change order migration, user runpython instead<commit_after>
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
def delete_force(apps, schema_editor):
# We can't import Infrastructure models directly as it may be a newer
# version than this migration expects. We use the historical version.
Blade = apps.get_model('signage', 'Blade')
for blade in Blade.objects.filter(deleted=True):
blade.delete()
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
migrations.RunPython(delete_force),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
]
|
# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
]
Change order migration, user runpython instead# Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
def delete_force(apps, schema_editor):
# We can't import Infrastructure models directly as it may be a newer
# version than this migration expects. We use the historical version.
Blade = apps.get_model('signage', 'Blade')
for blade in Blade.objects.filter(deleted=True):
blade.delete()
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
migrations.RunPython(delete_force),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
]
|
<commit_before># Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.RunSQL(sql=[("DELETE FROM geotrek.signage_blade WHERE deleted=TRUE;", )]),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
]
<commit_msg>Change order migration, user runpython instead<commit_after># Generated by Django 2.0.13 on 2020-04-23 12:55
from django.db import migrations, models
import django.db.models.deletion
def delete_force(apps, schema_editor):
# We can't import Infrastructure models directly as it may be a newer
# version than this migration expects. We use the historical version.
Blade = apps.get_model('signage', 'Blade')
for blade in Blade.objects.filter(deleted=True):
blade.delete()
class Migration(migrations.Migration):
dependencies = [
('signage', '0012_auto_20200406_1411'),
]
operations = [
migrations.AlterField(
model_name='line',
name='blade',
field=models.ForeignKey(on_delete=django.db.models.deletion.CASCADE, related_name='lines',
to='signage.Blade', verbose_name='Blade'),
),
migrations.RunPython(delete_force),
migrations.RemoveField(
model_name='blade',
name='deleted',
),
migrations.RemoveField(
model_name='blade',
name='structure',
),
migrations.RemoveField(
model_name='line',
name='structure',
),
]
|
b6d4baa9d30362a291567f078c1f93df7a63aeaa
|
waterbutler/providers/osfstorage/metadata.py
|
waterbutler/providers/osfstorage/metadata.py
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def full_path(self):
return self.raw.get('fullPath')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
Return full path if it exists for OSF
|
Return full path if it exists for OSF
|
Python
|
apache-2.0
|
cosenal/waterbutler,Ghalko/waterbutler,CenterForOpenScience/waterbutler,icereval/waterbutler,Johnetordoff/waterbutler,TomBaxter/waterbutler,rdhyee/waterbutler,felliott/waterbutler,RCOSDP/waterbutler,rafaeldelucena/waterbutler,chrisseto/waterbutler,hmoco/waterbutler,kwierman/waterbutler
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
Return full path if it exists for OSF
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def full_path(self):
return self.raw.get('fullPath')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
<commit_before>from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
<commit_msg>Return full path if it exists for OSF<commit_after>
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def full_path(self):
return self.raw.get('fullPath')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
Return full path if it exists for OSFfrom waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def full_path(self):
return self.raw.get('fullPath')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
<commit_before>from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
<commit_msg>Return full path if it exists for OSF<commit_after>from waterbutler.core import metadata
class BaseOsfStorageMetadata:
@property
def provider(self):
return 'osfstorage'
class OsfStorageFileMetadata(BaseOsfStorageMetadata, metadata.BaseFileMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
@property
def modified(self):
return self.raw.get('modified')
@property
def size(self):
return self.raw.get('size')
@property
def full_path(self):
return self.raw.get('fullPath')
@property
def content_type(self):
return None
@property
def extra(self):
return {
key: self.raw[key]
for key in
('version', 'downloads', 'fullPath')
if key in self.raw
}
class OsfStorageFolderMetadata(BaseOsfStorageMetadata, metadata.BaseFolderMetadata):
@property
def name(self):
return self.raw['name']
@property
def path(self):
return self.raw['path']
class OsfStorageRevisionMetadata(BaseOsfStorageMetadata, metadata.BaseFileRevisionMetadata):
@property
def modified(self):
return self.raw['date']
@property
def version_identifier(self):
return 'version'
@property
def version(self):
return str(self.raw['index'])
@property
def extra(self):
return {
'user': self.raw['user'],
'downloads': self.raw['downloads'],
}
|
f2fd526e08cc5576c651a7677c781c0c0bb7c94c
|
tests/test_jg.py
|
tests/test_jg.py
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
dot.render('t1.dot')
assert dot.source == FIXTURE_GRAPH
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
assert dot.source == FIXTURE_GRAPH
|
Fix test generating graph file
|
Fix test generating graph file
|
Python
|
bsd-2-clause
|
abele/jinja-graph
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
dot.render('t1.dot')
assert dot.source == FIXTURE_GRAPH
Fix test generating graph file
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
assert dot.source == FIXTURE_GRAPH
|
<commit_before>from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
dot.render('t1.dot')
assert dot.source == FIXTURE_GRAPH
<commit_msg>Fix test generating graph file<commit_after>
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
assert dot.source == FIXTURE_GRAPH
|
from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
dot.render('t1.dot')
assert dot.source == FIXTURE_GRAPH
Fix test generating graph filefrom jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
assert dot.source == FIXTURE_GRAPH
|
<commit_before>from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
dot.render('t1.dot')
assert dot.source == FIXTURE_GRAPH
<commit_msg>Fix test generating graph file<commit_after>from jg.__main__ import main, generate_template_graph
from mock import patch
FIXTURE_GRAPH = (
'digraph {\n'
'\t"snippets/sub/analytics.html"\n'
'\t"snippets/ga.html"\n'
'\t\t"snippets/ga.html" -> "snippets/sub/analytics.html"\n'
'\t"header.html"\n'
'\t"analytics.html"\n'
'\t"custom_index.html"\n'
'\t\t"custom_index.html" -> "index.html"\n'
'\t\t"custom_index.html" -> "snippets/ga.html"\n'
'\t"index.html"\n'
'\t\t"index.html" -> "header.html"\n'
'\t\t"index.html" -> "footer.html"\n'
'\t"footer.html"\n}')
def test_main_generates_graph_for_given_directory():
output_filename = 'graph.dot'
with patch('jg.__main__.write') as write:
exit_code = main(['./tests/fixtures', output_filename])
write.assert_called_with(FIXTURE_GRAPH, output_filename)
assert exit_code == 0
def test_parses_all_templates_in_given_root_directory():
dot = generate_template_graph(root_path='./tests/fixtures')
assert dot.source == FIXTURE_GRAPH
|
fe9a47f480b8db8de3b2b572f333e56497462ea2
|
Python/item15.py
|
Python/item15.py
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority2(num,pro):
found=[False]
def helper(x):
nonlocal found
if x in pro:
found[0]=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority2(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
Add the sort_priority2 for python2.
|
Add the sort_priority2 for python2.
|
Python
|
mit
|
Vayne-Lover/Effective
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
Add the sort_priority2 for python2.
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority2(num,pro):
found=[False]
def helper(x):
nonlocal found
if x in pro:
found[0]=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority2(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
<commit_before># -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
<commit_msg>Add the sort_priority2 for python2.<commit_after>
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority2(num,pro):
found=[False]
def helper(x):
nonlocal found
if x in pro:
found[0]=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority2(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
Add the sort_priority2 for python2.# -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority2(num,pro):
found=[False]
def helper(x):
nonlocal found
if x in pro:
found[0]=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority2(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
<commit_before># -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
<commit_msg>Add the sort_priority2 for python2.<commit_after># -*- coding: utf-8 -*-
def sort_priority(num,pro):
res=num[:]
def helper(x):
if x in pro:
return (0,x)
return (1,x)
res.sort(key=helper)
return res
def sort_priority2(num,pro):
found=[False]
def helper(x):
nonlocal found
if x in pro:
found[0]=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
def sort_priority3(num,pro):
found=False
def helper(x):
nonlocal found
if x in pro:
found=True
return (0,x)
return (1,x)
num.sort(key=helper)
return found
if __name__=='__main__':
numbers=[2,5,7,4,1,3,8,6]
group=[2,4,8]
print(sort_priority(numbers,group))
print(numbers)
print(sort_priority2(numbers,group))
print(numbers)
print(sort_priority3(numbers,group))
print(numbers)
|
8c7c2a65c9d9793e8ba4167ca629eda155f42c4e
|
mopidy/settings.py
|
mopidy/settings.py
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
Add logger name to log output
|
Add logger name to log output
|
Python
|
apache-2.0
|
quartz55/mopidy,bacontext/mopidy,mopidy/mopidy,mokieyue/mopidy,vrs01/mopidy,bencevans/mopidy,abarisain/mopidy,adamcik/mopidy,jmarsik/mopidy,tkem/mopidy,SuperStarPL/mopidy,jmarsik/mopidy,mokieyue/mopidy,ali/mopidy,dbrgn/mopidy,ZenithDK/mopidy,tkem/mopidy,priestd09/mopidy,adamcik/mopidy,woutervanwijk/mopidy,quartz55/mopidy,rawdlite/mopidy,liamw9534/mopidy,diandiankan/mopidy,adamcik/mopidy,hkariti/mopidy,rawdlite/mopidy,jodal/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,mokieyue/mopidy,vrs01/mopidy,jodal/mopidy,jmarsik/mopidy,tkem/mopidy,vrs01/mopidy,hkariti/mopidy,bencevans/mopidy,hkariti/mopidy,diandiankan/mopidy,priestd09/mopidy,abarisain/mopidy,SuperStarPL/mopidy,diandiankan/mopidy,swak/mopidy,bacontext/mopidy,dbrgn/mopidy,pacificIT/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,bacontext/mopidy,quartz55/mopidy,bencevans/mopidy,jcass77/mopidy,ZenithDK/mopidy,jcass77/mopidy,kingosticks/mopidy,ZenithDK/mopidy,ali/mopidy,jmarsik/mopidy,rawdlite/mopidy,swak/mopidy,pacificIT/mopidy,glogiotatidis/mopidy,dbrgn/mopidy,hkariti/mopidy,glogiotatidis/mopidy,ali/mopidy,liamw9534/mopidy,quartz55/mopidy,swak/mopidy,jodal/mopidy,bencevans/mopidy,swak/mopidy,vrs01/mopidy,mopidy/mopidy,dbrgn/mopidy,ZenithDK/mopidy,glogiotatidis/mopidy,mopidy/mopidy,mokieyue/mopidy,priestd09/mopidy,kingosticks/mopidy,rawdlite/mopidy,ali/mopidy,SuperStarPL/mopidy,kingosticks/mopidy,woutervanwijk/mopidy,jcass77/mopidy,tkem/mopidy,bacontext/mopidy,pacificIT/mopidy
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
Add logger name to log output
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
<commit_before>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
<commit_msg>Add logger name to log output<commit_after>
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
Add logger name to log outputCONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
<commit_before>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
<commit_msg>Add logger name to log output<commit_after>CONSOLE_LOG_FORMAT = u'%(levelname)-8s %(asctime)s %(name)s\n %(message)s'
MPD_LINE_ENCODING = u'utf-8'
MPD_LINE_TERMINATOR = u'\n'
MPD_SERVER_HOSTNAME = u'localhost'
MPD_SERVER_PORT = 6600
|
d4ea0bb24d79fa7900ffad4676a09cd4b830a4dc
|
parser_queuelog.py
|
parser_queuelog.py
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().split("\n")
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().splitlines()
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
Change split file lines by line
|
Change split file lines by line
|
Python
|
mit
|
roramirez/qpanel,roramirez/qpanel,roramirez/qpanel,skazancev/qpanel,skazancev/qpanel,skazancev/qpanel,skazancev/qpanel,roramirez/qpanel
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().split("\n")
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
Change split file lines by line
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().splitlines()
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
<commit_before># -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().split("\n")
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
<commit_msg>Change split file lines by line<commit_after>
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().splitlines()
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().split("\n")
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
Change split file lines by line# -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().splitlines()
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
<commit_before># -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().split("\n")
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
<commit_msg>Change split file lines by line<commit_after># -*- coding: utf-8 -*-
#
# Copyright (C) 2015-2016 Rodrigo Ramírez Norambuena <a@rodrigoramirez.com>
#
# Parse queue_log Asterisk file and add records into database.
#
from libs.qpanel import model
import click
import sys
@click.command()
@click.option('--file', default='/var/log/asterisk/queue_log',
help='Queue Log file.')
@click.option('--verbose', default=False)
def parse(file, verbose):
inserted, not_inserted = 0, 0
try:
with open(file) as fb:
print("Reading file %s ..." % file)
content = fb.read().splitlines()
except IOError:
print('File file %s not exits or not can read.' % file)
sys.exit(1)
for idx, line in enumerate(content):
record = line.split('|')
if len(record) < 4:
continue
if not exist_record(record) and insert_record(record):
inserted += 1
if verbose:
print ("Insert record ", record)
else:
if verbose:
print ("Not insert record ", record)
not_inserted += 1
print ("Insert record: %i\nNo inserted record: %i" %
(inserted, not_inserted))
def exist_record(record):
return model.queuelog_exists_record(record)
def insert_record(record):
return model.queuelog_insert(record)
if __name__ == '__main__':
parse()
|
0599e76db6c1eef495a608d7386601bbee3cfbc5
|
test/authinfo.py
|
test/authinfo.py
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Empty Result with HTTP GET
'''
def testEmptyInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
Add test for empty responses
|
Add test for empty responses
|
Python
|
mit
|
fbraem/mqweb,fbraem/mqweb,fbraem/mqweb
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
Add test for empty responses
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Empty Result with HTTP GET
'''
def testEmptyInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
<commit_before>import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
<commit_msg>Add test for empty responses<commit_after>
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Empty Result with HTTP GET
'''
def testEmptyInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
Add test for empty responsesimport unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Empty Result with HTTP GET
'''
def testEmptyInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
<commit_before>import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
<commit_msg>Add test for empty responses<commit_after>import unittest
from testbase import MQWebTest
'''
Test for AuthenticationInformationController
'''
class TestAuthInfoActions(MQWebTest):
'''
Test Inquire with HTTP GET
'''
def testInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr)
'''
Test Empty Result with HTTP GET
'''
def testEmptyInquire(self):
json = self.getJSON('/api/authinfo/inquire/' + self.qmgr + '/ZZZ*', True)
'''
Test Inquire with HTTP POST
'''
def testInquireJSON(self):
print
filter = { "AuthInfoName" : "*" }
json = self.appJSON('/api/authinfo/inquire/' + self.qmgr, filter)
suite = unittest.TestLoader().loadTestsFromTestCase(TestAuthInfoActions)
unittest.TextTestRunner(verbosity=2).run(suite)
|
d4a0a85673b5d61b82c65e77efcd6518da719952
|
pmxbot/__init__.py
|
pmxbot/__init__.py
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
Remove places default config. It doesn't appear to be used anywhere.
|
Remove places default config. It doesn't appear to be used anywhere.
|
Python
|
mit
|
yougov/pmxbot,yougov/pmxbot,yougov/pmxbot
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
Remove places default config. It doesn't appear to be used anywhere.
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
<commit_before># vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
<commit_msg>Remove places default config. It doesn't appear to be used anywhere.<commit_after>
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
Remove places default config. It doesn't appear to be used anywhere.# vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
<commit_before># vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
places=['London', 'Tokyo', 'New York'],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
<commit_msg>Remove places default config. It doesn't appear to be used anywhere.<commit_after># vim:ts=4:sw=4:noexpandtab
import socket
import logging as _logging
from .dictlib import ConfigDict
config = ConfigDict(
bot_nickname='pmxbot',
database='sqlite:pmxbot.sqlite',
server_host='localhost',
server_port=6667,
use_ssl=False,
password=None,
nickserv_password=None,
silent_bot=False,
log_channels=[],
other_channels=[],
librarypaste='http://paste.jaraco.com',
)
config['logs URL'] = 'http://' + socket.getfqdn()
config['log level'] = _logging.INFO
"The config object"
|
525cbab46570342098613ae591749b4cf5026453
|
tests/terrain.py
|
tests/terrain.py
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
# Create output directory if not exists
if not os.path.exists("test_output"):
os.makedirs("test_output")
|
Create the tests output directory automatically.
|
Create the tests output directory automatically.
|
Python
|
mit
|
gnott/elife-poa-xml-generation,gnott/elife-poa-xml-generation
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)Create the tests output directory automatically.
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
# Create output directory if not exists
if not os.path.exists("test_output"):
os.makedirs("test_output")
|
<commit_before>from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)<commit_msg>Create the tests output directory automatically.<commit_after>
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
# Create output directory if not exists
if not os.path.exists("test_output"):
os.makedirs("test_output")
|
from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)Create the tests output directory automatically.from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
# Create output directory if not exists
if not os.path.exists("test_output"):
os.makedirs("test_output")
|
<commit_before>from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)<commit_msg>Create the tests output directory automatically.<commit_after>from lettuce import world
import os
"""
Set world.basedir relative to this terrain.py file,
when running lettuce from this directory,
and add the directory it to the import path
"""
world.basedir = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
os.sys.path.insert(0,world.basedir)
world.basedir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
os.sys.path.insert(0,world.basedir)
# Create output directory if not exists
if not os.path.exists("test_output"):
os.makedirs("test_output")
|
8d09a8557433d95015010465b62f31ffe7b6fe2c
|
usb/shortener.py
|
usb/shortener.py
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = None
def get_short_id(self, number):
if self._hasher is None:
self._hasher = Hashids(self.secret, self.min_length)
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = Hashids(self.secret, self.min_length)
def get_short_id(self, number):
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
Move hasher creation to constructor
|
Move hasher creation to constructor
|
Python
|
mit
|
dizpers/usb
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = None
def get_short_id(self, number):
if self._hasher is None:
self._hasher = Hashids(self.secret, self.min_length)
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
Move hasher creation to constructor
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = Hashids(self.secret, self.min_length)
def get_short_id(self, number):
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
<commit_before>from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = None
def get_short_id(self, number):
if self._hasher is None:
self._hasher = Hashids(self.secret, self.min_length)
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
<commit_msg>Move hasher creation to constructor<commit_after>
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = Hashids(self.secret, self.min_length)
def get_short_id(self, number):
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = None
def get_short_id(self, number):
if self._hasher is None:
self._hasher = Hashids(self.secret, self.min_length)
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
Move hasher creation to constructorfrom hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = Hashids(self.secret, self.min_length)
def get_short_id(self, number):
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
<commit_before>from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = None
def get_short_id(self, number):
if self._hasher is None:
self._hasher = Hashids(self.secret, self.min_length)
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
<commit_msg>Move hasher creation to constructor<commit_after>from hashids import Hashids
class Shortener(object):
def __init__(self, secret, min_length, short_url_domain):
self.secret = secret
self.min_length = min_length
self.short_url_domain = short_url_domain
self._hasher = Hashids(self.secret, self.min_length)
def get_short_id(self, number):
return self._hasher.encode(number)
def get_short_url(self, short_id):
return f'http://{self.short_url_domain}/{short_id}'
|
b5ae6290382ef69f9d76c0494aee90f85bdf2c16
|
plugins/Views/SimpleView/__init__.py
|
plugins/Views/SimpleView/__init__.py
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"type": "view",
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
Fix plug-in type and description key
|
Fix plug-in type and description key
't Was a typo.
Contributes to issue CURA-1190.
|
Python
|
agpl-3.0
|
onitake/Uranium,onitake/Uranium
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
Fix plug-in type and description key
't Was a typo.
Contributes to issue CURA-1190.
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"type": "view",
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
<commit_before># Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
<commit_msg>Fix plug-in type and description key
't Was a typo.
Contributes to issue CURA-1190.<commit_after>
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"type": "view",
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
Fix plug-in type and description key
't Was a typo.
Contributes to issue CURA-1190.# Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"type": "view",
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
<commit_before># Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"decription": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
<commit_msg>Fix plug-in type and description key
't Was a typo.
Contributes to issue CURA-1190.<commit_after># Copyright (c) 2015 Ultimaker B.V.
# Uranium is released under the terms of the AGPLv3 or higher.
from . import SimpleView
from UM.i18n import i18nCatalog
i18n_catalog = i18nCatalog("uranium")
def getMetaData():
return {
"type": "view",
"plugin": {
"name": i18n_catalog.i18nc("@label", "Simple View"),
"author": "Ultimaker",
"version": "1.0",
"description": i18n_catalog.i18nc("@info:whatsthis", "Provides a simple solid mesh view."),
"api": 2
},
"view": {
"name": i18n_catalog.i18nc("@item:inmenu", "Simple"),
"visible": False
}
}
def register(app):
return { "view": SimpleView.SimpleView() }
|
35c7d7816c3c441286519658a3426a5f03aca284
|
plugins/check_pinned/check_pinned.py
|
plugins/check_pinned/check_pinned.py
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
# Can we send a message to the Slackbot with the UserID
# (which we have) and have the slackbot post to the user?
|
Add note about potential work around
|
Add note about potential work around
The bot itself cannot send a message only on presence_change actions
We may be able to contact the slack bot to send a message in our stead instead.
|
Python
|
mit
|
pyamanak/oithdbot
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
Add note about potential work around
The bot itself cannot send a message only on presence_change actions
We may be able to contact the slack bot to send a message in our stead instead.
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
# Can we send a message to the Slackbot with the UserID
# (which we have) and have the slackbot post to the user?
|
<commit_before>from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
<commit_msg>Add note about potential work around
The bot itself cannot send a message only on presence_change actions
We may be able to contact the slack bot to send a message in our stead instead.<commit_after>
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
# Can we send a message to the Slackbot with the UserID
# (which we have) and have the slackbot post to the user?
|
from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
Add note about potential work around
The bot itself cannot send a message only on presence_change actions
We may be able to contact the slack bot to send a message in our stead instead.from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
# Can we send a message to the Slackbot with the UserID
# (which we have) and have the slackbot post to the user?
|
<commit_before>from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
<commit_msg>Add note about potential work around
The bot itself cannot send a message only on presence_change actions
We may be able to contact the slack bot to send a message in our stead instead.<commit_after>from __future__ import unicode_literals
# don't convert to ascii in py2.7 when creating string to return
crontable = []
outputs = []
# Catch all the events
def catch_all(data):
print(data)
# Only handles when a user becomes active
def process_presence_change(data):
print("PRESENCE CHANGE")
# While we can respond to presence change events,
# we cannot actually send a message to a channel as
# the data structure does not contain a channel ID
if (data["presence"].startswith("active")):
print("IS ACTIVE")
# Can we send a message to the Slackbot with the UserID
# (which we have) and have the slackbot post to the user?
|
e4ee7034291fbeda48efa0d1c617be8a20eb49bd
|
algorithms/python/496_next_greater_element.py
|
algorithms/python/496_next_greater_element.py
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
|
Add another solution for 496 next greater element
|
Add another solution for 496 next greater element
|
Python
|
mit
|
ruichao-factual/leetcode
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
Add another solution for 496 next greater element
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
|
<commit_before>class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
<commit_msg>Add another solution for 496 next greater element<commit_after>
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
|
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
Add another solution for 496 next greater elementclass Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
|
<commit_before>class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
<commit_msg>Add another solution for 496 next greater element<commit_after>class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
results = []
for findNum in findNums:
index = nums.index(findNum)
result = index + 1
for candidate in nums[index + 1:]:
if candidate > findNum:
results.append(candidate)
break
else:
result += 1
if result >= len(nums):
results.append(-1)
return results
# Solution 2
class Solution(object):
def nextGreaterElement(self, findNums, nums):
"""
:type findNums: List[int]
:type nums: List[int]
:rtype: List[int]
"""
result_hash = {}
stack = []
for num in nums:
while stack and num > stack[-1]:
result_hash[stack.pop()] = num
stack.append(num)
return [result_hash.get(x, -1) for x in findNums]
|
c480c2058d7f5d2e9bfe2e1a6f0a318eb0d49a4b
|
tangled/website/resources.py
|
tangled/website/resources.py
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template_name='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
Rename template_name @config arg to template
|
Rename template_name @config arg to template
|
Python
|
mit
|
TangledWeb/tangled.website
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template_name='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
Rename template_name @config arg to template
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
<commit_before>from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template_name='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
<commit_msg>Rename template_name @config arg to template<commit_after>
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template_name='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
Rename template_name @config arg to templatefrom tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
<commit_before>from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template_name='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
<commit_msg>Rename template_name @config arg to template<commit_after>from tangled.web import Resource, config
from tangled.site.resources.entry import Entry
class Docs(Entry):
@config('text/html', template='tangled.website:templates/docs.mako')
def GET(self):
static_dirs = self.app.get_all('static_directory', as_dict=True)
links = []
for prefix, dir_app in static_dirs.items():
if prefix[0] == 'docs':
links.append({
'href': '/'.join(prefix),
'text': prefix[1],
})
self.urlvars['id'] = 'docs'
data = super().GET()
data['links'] = sorted(links, key=lambda i: i['text'])
return data
|
428371a67a7c742e239d96914a96558171f9f59e
|
test/requests/test-website.py
|
test/requests/test-website.py
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
import requests as req
import sys
print "Mechanical Rob firing up..."
if len(sys.argv)<1:
raise "Problem with arguments"
url = sys.argv[1]
print url
r = req.get(url)
print r
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
from __future__ import print_function
from link_checker import check_links
import argparse
print("Mechanical Rob firing up...")
def run_all(args_obj, parser):
print("")
print("Running all tests.")
check_links(args_obj, parser)
# TODO: Add other functions as they are created.
def print_help(args_obj, parser):
print(parser.format_help())
def dummy(args_obj, parser):
print("Not implemented yet.")
desc = """
This is Mechanical-Rob - an automated web server tester for
Genenetwork.org
"""
parser = argparse.ArgumentParser(description=desc)
parser.add_argument("-d", "--database", metavar="DB", type=str
, default="db_webqtl_s"
, help="Use database (default db_webqtl_s)")
parser.add_argument("host", metavar="HOST", type=str
, default="http://localhost:5003"
, help="The url to the web server")
parser.add_argument("-a", "--all", dest="accumulate", action="store_const"
, const=run_all, default=print_help
, help="Runs all tests.")
parser.add_argument("-l", "--link-checker", dest="accumulate"
, action='store_const', const=check_links, default=print_help
, help="Checks for dead links.")
# parser.add_argument("-n", "--navigation", dest="accumulate"
# , action="store_const", const=check_navigation, default=print_help
# , help="Checks for navigation.")
# parser.add_argument("-m", "--mapping", dest="accumulate"
# , action="store_const", const=check_mapping, default=print_help
# , help="Checks for mapping.")
# parser.add_argument("-s", "--skip-broken", dest="accumulate"
# , action="store_const", const=dummy, default=print_help
# , help="Skip tests that are known to be broken.")
args = parser.parse_args()
# print("The arguments object: ", args)
args.accumulate(args, parser)
|
Use argparse to handle arguments
|
Use argparse to handle arguments
* Use argparse to handle commandline arguments.
* Create initial layout of how the code might end up - lots of the code
is currently commented out.
|
Python
|
agpl-3.0
|
zsloan/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,genenetwork/genenetwork2,pjotrp/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,pjotrp/genenetwork2,genenetwork/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,DannyArends/genenetwork2,pjotrp/genenetwork2,zsloan/genenetwork2,zsloan/genenetwork2
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
import requests as req
import sys
print "Mechanical Rob firing up..."
if len(sys.argv)<1:
raise "Problem with arguments"
url = sys.argv[1]
print url
r = req.get(url)
print r
Use argparse to handle arguments
* Use argparse to handle commandline arguments.
* Create initial layout of how the code might end up - lots of the code
is currently commented out.
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
from __future__ import print_function
from link_checker import check_links
import argparse
print("Mechanical Rob firing up...")
def run_all(args_obj, parser):
print("")
print("Running all tests.")
check_links(args_obj, parser)
# TODO: Add other functions as they are created.
def print_help(args_obj, parser):
print(parser.format_help())
def dummy(args_obj, parser):
print("Not implemented yet.")
desc = """
This is Mechanical-Rob - an automated web server tester for
Genenetwork.org
"""
parser = argparse.ArgumentParser(description=desc)
parser.add_argument("-d", "--database", metavar="DB", type=str
, default="db_webqtl_s"
, help="Use database (default db_webqtl_s)")
parser.add_argument("host", metavar="HOST", type=str
, default="http://localhost:5003"
, help="The url to the web server")
parser.add_argument("-a", "--all", dest="accumulate", action="store_const"
, const=run_all, default=print_help
, help="Runs all tests.")
parser.add_argument("-l", "--link-checker", dest="accumulate"
, action='store_const', const=check_links, default=print_help
, help="Checks for dead links.")
# parser.add_argument("-n", "--navigation", dest="accumulate"
# , action="store_const", const=check_navigation, default=print_help
# , help="Checks for navigation.")
# parser.add_argument("-m", "--mapping", dest="accumulate"
# , action="store_const", const=check_mapping, default=print_help
# , help="Checks for mapping.")
# parser.add_argument("-s", "--skip-broken", dest="accumulate"
# , action="store_const", const=dummy, default=print_help
# , help="Skip tests that are known to be broken.")
args = parser.parse_args()
# print("The arguments object: ", args)
args.accumulate(args, parser)
|
<commit_before># Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
import requests as req
import sys
print "Mechanical Rob firing up..."
if len(sys.argv)<1:
raise "Problem with arguments"
url = sys.argv[1]
print url
r = req.get(url)
print r
<commit_msg>Use argparse to handle arguments
* Use argparse to handle commandline arguments.
* Create initial layout of how the code might end up - lots of the code
is currently commented out.<commit_after>
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
from __future__ import print_function
from link_checker import check_links
import argparse
print("Mechanical Rob firing up...")
def run_all(args_obj, parser):
print("")
print("Running all tests.")
check_links(args_obj, parser)
# TODO: Add other functions as they are created.
def print_help(args_obj, parser):
print(parser.format_help())
def dummy(args_obj, parser):
print("Not implemented yet.")
desc = """
This is Mechanical-Rob - an automated web server tester for
Genenetwork.org
"""
parser = argparse.ArgumentParser(description=desc)
parser.add_argument("-d", "--database", metavar="DB", type=str
, default="db_webqtl_s"
, help="Use database (default db_webqtl_s)")
parser.add_argument("host", metavar="HOST", type=str
, default="http://localhost:5003"
, help="The url to the web server")
parser.add_argument("-a", "--all", dest="accumulate", action="store_const"
, const=run_all, default=print_help
, help="Runs all tests.")
parser.add_argument("-l", "--link-checker", dest="accumulate"
, action='store_const', const=check_links, default=print_help
, help="Checks for dead links.")
# parser.add_argument("-n", "--navigation", dest="accumulate"
# , action="store_const", const=check_navigation, default=print_help
# , help="Checks for navigation.")
# parser.add_argument("-m", "--mapping", dest="accumulate"
# , action="store_const", const=check_mapping, default=print_help
# , help="Checks for mapping.")
# parser.add_argument("-s", "--skip-broken", dest="accumulate"
# , action="store_const", const=dummy, default=print_help
# , help="Skip tests that are known to be broken.")
args = parser.parse_args()
# print("The arguments object: ", args)
args.accumulate(args, parser)
|
# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
import requests as req
import sys
print "Mechanical Rob firing up..."
if len(sys.argv)<1:
raise "Problem with arguments"
url = sys.argv[1]
print url
r = req.get(url)
print r
Use argparse to handle arguments
* Use argparse to handle commandline arguments.
* Create initial layout of how the code might end up - lots of the code
is currently commented out.# Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
from __future__ import print_function
from link_checker import check_links
import argparse
print("Mechanical Rob firing up...")
def run_all(args_obj, parser):
print("")
print("Running all tests.")
check_links(args_obj, parser)
# TODO: Add other functions as they are created.
def print_help(args_obj, parser):
print(parser.format_help())
def dummy(args_obj, parser):
print("Not implemented yet.")
desc = """
This is Mechanical-Rob - an automated web server tester for
Genenetwork.org
"""
parser = argparse.ArgumentParser(description=desc)
parser.add_argument("-d", "--database", metavar="DB", type=str
, default="db_webqtl_s"
, help="Use database (default db_webqtl_s)")
parser.add_argument("host", metavar="HOST", type=str
, default="http://localhost:5003"
, help="The url to the web server")
parser.add_argument("-a", "--all", dest="accumulate", action="store_const"
, const=run_all, default=print_help
, help="Runs all tests.")
parser.add_argument("-l", "--link-checker", dest="accumulate"
, action='store_const', const=check_links, default=print_help
, help="Checks for dead links.")
# parser.add_argument("-n", "--navigation", dest="accumulate"
# , action="store_const", const=check_navigation, default=print_help
# , help="Checks for navigation.")
# parser.add_argument("-m", "--mapping", dest="accumulate"
# , action="store_const", const=check_mapping, default=print_help
# , help="Checks for mapping.")
# parser.add_argument("-s", "--skip-broken", dest="accumulate"
# , action="store_const", const=dummy, default=print_help
# , help="Skip tests that are known to be broken.")
args = parser.parse_args()
# print("The arguments object: ", args)
args.accumulate(args, parser)
|
<commit_before># Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
import requests as req
import sys
print "Mechanical Rob firing up..."
if len(sys.argv)<1:
raise "Problem with arguments"
url = sys.argv[1]
print url
r = req.get(url)
print r
<commit_msg>Use argparse to handle arguments
* Use argparse to handle commandline arguments.
* Create initial layout of how the code might end up - lots of the code
is currently commented out.<commit_after># Run with something like
#
# env GN2_PROFILE=/home/wrk/opt/gn-latest ./bin/genenetwork2 ./etc/default_settings.py -c ../test/requests/test-website.py http://localhost:5003
#
# Mostly to pick up the Guix GN2_PROFILE and python modules
from __future__ import print_function
from link_checker import check_links
import argparse
print("Mechanical Rob firing up...")
def run_all(args_obj, parser):
print("")
print("Running all tests.")
check_links(args_obj, parser)
# TODO: Add other functions as they are created.
def print_help(args_obj, parser):
print(parser.format_help())
def dummy(args_obj, parser):
print("Not implemented yet.")
desc = """
This is Mechanical-Rob - an automated web server tester for
Genenetwork.org
"""
parser = argparse.ArgumentParser(description=desc)
parser.add_argument("-d", "--database", metavar="DB", type=str
, default="db_webqtl_s"
, help="Use database (default db_webqtl_s)")
parser.add_argument("host", metavar="HOST", type=str
, default="http://localhost:5003"
, help="The url to the web server")
parser.add_argument("-a", "--all", dest="accumulate", action="store_const"
, const=run_all, default=print_help
, help="Runs all tests.")
parser.add_argument("-l", "--link-checker", dest="accumulate"
, action='store_const', const=check_links, default=print_help
, help="Checks for dead links.")
# parser.add_argument("-n", "--navigation", dest="accumulate"
# , action="store_const", const=check_navigation, default=print_help
# , help="Checks for navigation.")
# parser.add_argument("-m", "--mapping", dest="accumulate"
# , action="store_const", const=check_mapping, default=print_help
# , help="Checks for mapping.")
# parser.add_argument("-s", "--skip-broken", dest="accumulate"
# , action="store_const", const=dummy, default=print_help
# , help="Skip tests that are known to be broken.")
args = parser.parse_args()
# print("The arguments object: ", args)
args.accumulate(args, parser)
|
e5daa53aab94360c2e06a6cb608c4992b25becc6
|
test/helpers.py
|
test/helpers.py
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(1)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(1)
f.data = data
f.flags = flags
return f
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None, stream_id=1):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(stream_id)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None, stream_id=1):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(stream_id)
f.data = data
f.flags = flags
return f
|
Allow stream ids != 1 in frame factory.
|
Allow stream ids != 1 in frame factory.
|
Python
|
mit
|
Kriechi/hyper-h2,mhils/hyper-h2,bhavishyagopesh/hyper-h2,Kriechi/hyper-h2,vladmunteanu/hyper-h2,python-hyper/hyper-h2,python-hyper/hyper-h2,vladmunteanu/hyper-h2
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(1)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(1)
f.data = data
f.flags = flags
return f
Allow stream ids != 1 in frame factory.
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None, stream_id=1):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(stream_id)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None, stream_id=1):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(stream_id)
f.data = data
f.flags = flags
return f
|
<commit_before># -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(1)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(1)
f.data = data
f.flags = flags
return f
<commit_msg>Allow stream ids != 1 in frame factory.<commit_after>
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None, stream_id=1):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(stream_id)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None, stream_id=1):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(stream_id)
f.data = data
f.flags = flags
return f
|
# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(1)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(1)
f.data = data
f.flags = flags
return f
Allow stream ids != 1 in frame factory.# -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None, stream_id=1):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(stream_id)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None, stream_id=1):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(stream_id)
f.data = data
f.flags = flags
return f
|
<commit_before># -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(1)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(1)
f.data = data
f.flags = flags
return f
<commit_msg>Allow stream ids != 1 in frame factory.<commit_after># -*- coding: utf-8 -*-
"""
helpers
~~~~~~~
This module contains helpers for the h2 tests.
"""
from hyperframe.frame import HeadersFrame, DataFrame
from hpack.hpack import Encoder
class FrameFactory(object):
"""
A class containing lots of helper methods and state to build frames. This
allows test cases to easily build correct HTTP/2 frames to feed to
hyper-h2.
"""
def __init__(self):
self.encoder = Encoder()
def preamble(self):
return b'PRI * HTTP/2.0\r\n\r\nSM\r\n\r\n'
def build_headers_frame(self, headers, flags=None, stream_id=1):
"""
Builds a single valid headers frame out of the contained headers.
"""
f = HeadersFrame(stream_id)
f.data = self.encoder.encode(headers)
f.flags.add('END_HEADERS')
if flags:
f.flags.update(flags)
return f
def build_data_frame(self, data, flags=None, stream_id=1):
"""
Builds a single data frame out of a chunk of data.
"""
flags = set(flags) if flags is not None else set()
f = DataFrame(stream_id)
f.data = data
f.flags = flags
return f
|
1509336a27d80eae68e56cfa776bd8342221297f
|
tests/scoring_engine/engine/test_basic_check.py
|
tests/scoring_engine/engine/test_basic_check.py
|
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
|
import pytest
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
def test_check_no_properties(self):
check = BasicCheck(self.environment)
check.required_properties = ['testparam']
with pytest.raises(LookupError):
check.set_properties()
|
Add test for incorrect check properties
|
Add test for incorrect check properties
|
Python
|
mit
|
pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine,pwnbus/scoring_engine
|
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
Add test for incorrect check properties
|
import pytest
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
def test_check_no_properties(self):
check = BasicCheck(self.environment)
check.required_properties = ['testparam']
with pytest.raises(LookupError):
check.set_properties()
|
<commit_before>from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
<commit_msg>Add test for incorrect check properties<commit_after>
|
import pytest
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
def test_check_no_properties(self):
check = BasicCheck(self.environment)
check.required_properties = ['testparam']
with pytest.raises(LookupError):
check.set_properties()
|
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
Add test for incorrect check propertiesimport pytest
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
def test_check_no_properties(self):
check = BasicCheck(self.environment)
check.required_properties = ['testparam']
with pytest.raises(LookupError):
check.set_properties()
|
<commit_before>from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
<commit_msg>Add test for incorrect check properties<commit_after>import pytest
from scoring_engine.engine.basic_check import BasicCheck
from scoring_engine.models.service import Service
from scoring_engine.models.environment import Environment
from scoring_engine.models.account import Account
from tests.scoring_engine.unit_test import UnitTest
class TestBasicCheck(UnitTest):
def setup(self):
super(TestBasicCheck, self).setup()
self.service = Service(name="Example Service", check_name="ICMP IPv4 Check", host='127.0.0.1')
self.environment = Environment(matching_regex='*', service=self.service)
def test_init(self):
check = BasicCheck(self.environment)
assert check.environment == self.environment
assert check.required_properties == []
def test_get_host(self):
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.host == '127.0.0.1'
def test_get_random_account(self):
self.session.add(Account(username='pwnbus', password='pass', service=self.service))
self.session.add(self.service)
self.session.add(self.environment)
self.session.commit()
check = BasicCheck(self.environment)
assert check.get_random_account().username == 'pwnbus'
def test_check_no_properties(self):
check = BasicCheck(self.environment)
check.required_properties = ['testparam']
with pytest.raises(LookupError):
check.set_properties()
|
8b5ccf93fbac8929ecfc185d7407a79b1e890bde
|
project_template/project_settings.py
|
project_template/project_settings.py
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
Use GLAMkit settings in default project template.
|
Use GLAMkit settings in default project template.
|
Python
|
mit
|
ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit,ic-labs/django-icekit
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
Use GLAMkit settings in default project template.
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
<commit_before># Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
<commit_msg>Use GLAMkit settings in default project template.<commit_after>
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
Use GLAMkit settings in default project template.# Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
<commit_before># Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.icekit import * # icekit, glamkit
# Override the default ICEkit settings to form project settings.
<commit_msg>Use GLAMkit settings in default project template.<commit_after># Do not commit secrets to VCS.
# Local environment variables will be loaded from `.env.local`.
# Additional environment variables will be loaded from `.env.$DOTENV`.
# Local settings will be imported from `icekit_settings_local.py`
from icekit.project.settings.glamkit import * # glamkit, icekit
# Override the default ICEkit settings to form project settings.
|
ea20424f2645f1157c0a7ea76b32a4f834c1e27b
|
pythonect/internal/__init__.py
|
pythonect/internal/__init__.py
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
Delete blank line at end of file
|
Delete blank line at end of file
|
Python
|
bsd-3-clause
|
ikotler/pythonect,fr34k8/pythonect,fr34k8/pythonect,TOSPIO/pythonect,ikotler/pythonect
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
Delete blank line at end of file
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
<commit_before># Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
<commit_msg>Delete blank line at end of file<commit_after>
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
Delete blank line at end of file# Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
<commit_before># Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
<commit_msg>Delete blank line at end of file<commit_after># Copyright (c) 2012-2013, Itzik Kotler
# All rights reserved.
#
# Redistribution and use in source and binary forms, with or without
# modification, are permitted provided that the following conditions are met:
#
# * Redistributions of source code must retain the above copyright
# notice, this list of conditions and the following disclaimer.
#
# * Redistributions in binary form must reproduce the above copyright
# notice, this list of conditions and the following disclaimer in the
# documentation and/or other materials provided with the distribution.
#
# * Neither the name of the author nor the names of its contributors may
# be used to endorse or promote products derived from this software without
# specific prior written permission.
#
# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND
# ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
# WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
# DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR ANY
# DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
# (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
# LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND
# ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
# SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
6cfc784ce3136cbec8c88948f4d6b45f9070b91b
|
pyqode/__init__.py
|
pyqode/__init__.py
|
# -*- coding: utf-8 -*-
"""
pyQode is a code editor framework for python qt applications.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
# -*- coding: utf-8 -*-
"""
pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide)
pyQode is a **namespace package**.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
Fix pyqode main docstring (uniformised with other pyqode namespace packages)
|
Fix pyqode main docstring (uniformised with other pyqode namespace packages)
|
Python
|
mit
|
mmolero/pyqode.python,pyQode/pyqode.python,zwadar/pyqode.python,pyQode/pyqode.python
|
# -*- coding: utf-8 -*-
"""
pyQode is a code editor framework for python qt applications.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
Fix pyqode main docstring (uniformised with other pyqode namespace packages)
|
# -*- coding: utf-8 -*-
"""
pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide)
pyQode is a **namespace package**.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
<commit_before># -*- coding: utf-8 -*-
"""
pyQode is a code editor framework for python qt applications.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
<commit_msg>Fix pyqode main docstring (uniformised with other pyqode namespace packages)<commit_after>
|
# -*- coding: utf-8 -*-
"""
pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide)
pyQode is a **namespace package**.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
# -*- coding: utf-8 -*-
"""
pyQode is a code editor framework for python qt applications.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
Fix pyqode main docstring (uniformised with other pyqode namespace packages)# -*- coding: utf-8 -*-
"""
pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide)
pyQode is a **namespace package**.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
<commit_before># -*- coding: utf-8 -*-
"""
pyQode is a code editor framework for python qt applications.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
<commit_msg>Fix pyqode main docstring (uniformised with other pyqode namespace packages)<commit_after># -*- coding: utf-8 -*-
"""
pyQode is a source code editor widget for Python Qt (PyQt5/PyQt4/PySide)
pyQode is a **namespace package**.
"""
import pkg_resources
pkg_resources.declare_namespace(__name__)
|
226e8c322670a310fcfb9eb95d9d59838bbac3d3
|
refcollections/admin_custom.py
|
refcollections/admin_custom.py
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.auth.admin import UserAdmin
from django.contrib.auth.models import User
shells_admin.register(User, UserAdmin)
|
Add User back into admin
|
Add User back into admin
|
Python
|
bsd-3-clause
|
uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections,uq-eresearch/archaeology-reference-collections
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
Add User back into admin
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.auth.admin import UserAdmin
from django.contrib.auth.models import User
shells_admin.register(User, UserAdmin)
|
<commit_before>from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
<commit_msg>Add User back into admin<commit_after>
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.auth.admin import UserAdmin
from django.contrib.auth.models import User
shells_admin.register(User, UserAdmin)
|
from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
Add User back into adminfrom django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.auth.admin import UserAdmin
from django.contrib.auth.models import User
shells_admin.register(User, UserAdmin)
|
<commit_before>from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
<commit_msg>Add User back into admin<commit_after>from django.contrib.admin.sites import AdminSite
from django.conf.urls.defaults import patterns, url
from shells.admin_views import ShellsImagesUploader, upload_species_spreadsheet
class ShellsAdmin(AdminSite):
def get_urls(self):
urls = super(ShellsAdmin, self).get_urls()
my_urls = patterns('',
url('upload_images/',
self.admin_view(ShellsImagesUploader.as_view()),
name="upload-images"),
url(r'^upload/', self.admin_view(upload_species_spreadsheet),
name='upload_species_spreadsheet'),
)
return my_urls + urls
shells_admin = ShellsAdmin()
from shells.admin import SpeciesAdmin, SpecimenAdmin, SpeciesRepresentationAdmin
from shells.models import Species, Specimen, SpeciesRepresentation
shells_admin.register(Species, SpeciesAdmin)
shells_admin.register(Specimen, SpecimenAdmin)
shells_admin.register(SpeciesRepresentation, SpeciesRepresentationAdmin)
from django.contrib.auth.admin import UserAdmin
from django.contrib.auth.models import User
shells_admin.register(User, UserAdmin)
|
52d38e360b14fcfad01f87ff1e9ca5db27004877
|
src/comms/admin.py
|
src/comms/admin.py
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions")
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage')
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
|
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
|
Python
|
bsd-3-clause
|
ypwalter/evennia,TheTypoMaster/evennia,TheTypoMaster/evennia,mrkulk/text-world,mrkulk/text-world,titeuf87/evennia,ergodicbreak/evennia,mrkulk/text-world,feend78/evennia,shollen/evennia,jamesbeebop/evennia,shollen/evennia,feend78/evennia,ergodicbreak/evennia,feend78/evennia,titeuf87/evennia,mrkulk/text-world,jamesbeebop/evennia,emergebtc/evennia,ypwalter/evennia,titeuf87/evennia,TheTypoMaster/evennia,emergebtc/evennia,titeuf87/evennia,ypwalter/evennia,feend78/evennia,emergebtc/evennia,jamesbeebop/evennia,ergodicbreak/evennia
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions")
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage')
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
<commit_before>#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions")
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
<commit_msg>Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.<commit_after>
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage')
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions")
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage')
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
<commit_before>#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage', "db_subscriptions")
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
<commit_msg>Remove unsupport M2M field in channelAdmin handler. Removes traceback when DEBUG=True.<commit_after>#
# This sets up how models are displayed
# in the web admin interface.
#
from django.contrib import admin
from src.comms.models import ChannelDB
class MsgAdmin(admin.ModelAdmin):
list_display = ('id', 'db_date_sent', 'db_sender', 'db_receivers',
'db_channels', 'db_message', 'db_lock_storage')
list_display_links = ("id",)
ordering = ["db_date_sent", 'db_sender', 'db_receivers', 'db_channels']
#readonly_fields = ['db_message', 'db_sender', 'db_receivers', 'db_channels']
search_fields = ['id', '^db_date_sent', '^db_message']
save_as = True
save_on_top = True
list_select_related = True
#admin.site.register(Msg, MsgAdmin)
class ChannelAdmin(admin.ModelAdmin):
list_display = ('id', 'db_key', 'db_lock_storage')
list_display_links = ("id", 'db_key')
ordering = ["db_key"]
search_fields = ['id', 'db_key', 'db_aliases']
save_as = True
save_on_top = True
list_select_related = True
fieldsets = (
(None, {'fields': (('db_key',), 'db_lock_storage')}),
)
admin.site.register(ChannelDB, ChannelAdmin)
|
21b6a5573190848b93de930b9d41e1ac766c18bc
|
src/epiweb/urls.py
|
src/epiweb/urls.py
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
)
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
(r'^accounts/', include('registration.backends.default.urls')),
)
|
Add user registration URLs. Use what django-registration provides for the moment.
|
Add user registration URLs. Use what django-registration provides for the moment.
|
Python
|
agpl-3.0
|
ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
)
Add user registration URLs. Use what django-registration provides for the moment.
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
(r'^accounts/', include('registration.backends.default.urls')),
)
|
<commit_before>from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Add user registration URLs. Use what django-registration provides for the moment.<commit_after>
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
(r'^accounts/', include('registration.backends.default.urls')),
)
|
from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
)
Add user registration URLs. Use what django-registration provides for the moment.from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
(r'^accounts/', include('registration.backends.default.urls')),
)
|
<commit_before>from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
)
<commit_msg>Add user registration URLs. Use what django-registration provides for the moment.<commit_after>from django.conf.urls.defaults import *
# Uncomment the next two lines to enable the admin:
from django.contrib import admin
admin.autodiscover()
import epiweb.apps.survey.urls
urlpatterns = patterns('',
# Example:
# (r'^epiweb/', include('epiweb.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
(r'^admin/', include(admin.site.urls)),
(r'^accounts/', include('registration.backends.default.urls')),
)
|
017de01e8a1ec8f49069cf546e89652b4ddb8e39
|
tests/test_create_template.py
|
tests/test_create_template.py
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
@pytest.fixture
def output_dir(tmpdir):
return str(tmpdir.mkdir('output'))
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
Remove output_dir fixture from test
|
Remove output_dir fixture from test
|
Python
|
mit
|
pytest-dev/cookiecutter-pytest-plugin
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
@pytest.fixture
def output_dir(tmpdir):
return str(tmpdir.mkdir('output'))
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
Remove output_dir fixture from test
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
<commit_before># -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
@pytest.fixture
def output_dir(tmpdir):
return str(tmpdir.mkdir('output'))
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
<commit_msg>Remove output_dir fixture from test<commit_after>
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
@pytest.fixture
def output_dir(tmpdir):
return str(tmpdir.mkdir('output'))
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
Remove output_dir fixture from test# -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
<commit_before># -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
@pytest.fixture
def output_dir(tmpdir):
return str(tmpdir.mkdir('output'))
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
<commit_msg>Remove output_dir fixture from test<commit_after># -*- coding: utf-8 -*-
"""
test_create_template
--------------------
"""
import os
import pytest
import subprocess
def run_tox(plugin):
"""Run the tox suite of the newly created plugin."""
try:
subprocess.check_call([
'tox',
plugin,
'-c', os.path.join(plugin, 'tox.ini'),
'-e', 'py'
])
except subprocess.CalledProcessError as e:
pytest.fail(e)
def test_run_cookiecutter_and_plugin_tests(cookies):
"""Create a new plugin via cookiecutter and run its tests."""
result = cookies.bake()
assert result.project.basename == 'pytest-foobar'
assert result.project.isdir()
run_tox(str(result.project))
|
a6e9438cd7aa205fb1ae1f20d192b9139bcd7f95
|
thecure/game.py
|
thecure/game.py
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((800, 600))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((1024, 768))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
Increase the resolution to 1024x768.
|
Increase the resolution to 1024x768.
This will make it easier to see things like the house. 800x600 is just
too small.
|
Python
|
mit
|
chipx86/the-cure
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((800, 600))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
Increase the resolution to 1024x768.
This will make it easier to see things like the house. 800x600 is just
too small.
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((1024, 768))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
<commit_before>import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((800, 600))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
<commit_msg>Increase the resolution to 1024x768.
This will make it easier to see things like the house. 800x600 is just
too small.<commit_after>
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((1024, 768))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((800, 600))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
Increase the resolution to 1024x768.
This will make it easier to see things like the house. 800x600 is just
too small.import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((1024, 768))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
<commit_before>import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((800, 600))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
<commit_msg>Increase the resolution to 1024x768.
This will make it easier to see things like the house. 800x600 is just
too small.<commit_after>import pygame
from pygame.locals import *
from thecure.engine import TheCureEngine
def main():
pygame.init()
version = pygame.__version__.split('.')
if int(version[0]) <= 1 and int(version[1]) < 9:
print 'This game requires pygame 1.9 or higher.'
return
screen = pygame.display.set_mode((1024, 768))
pygame.display.set_caption('The Cure')
engine = TheCureEngine(screen)
engine.run()
pygame.quit()
|
18b0ddbbca429778a70f1e9b7f7d5140eb88d68f
|
tests/test_fs.py
|
tests/test_fs.py
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
def test_cmp():
assert Path("/a/b") < Path("/a/c")
assert Path("/a/c") > Path("/a/b")
assert Path("/a/2") < Path("/b/1")
|
Add test to Path compare.
|
Add test to Path compare.
|
Python
|
mit
|
andrewguy9/farmfs,andrewguy9/farmfs
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
Add test to Path compare.
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
def test_cmp():
assert Path("/a/b") < Path("/a/c")
assert Path("/a/c") > Path("/a/b")
assert Path("/a/2") < Path("/b/1")
|
<commit_before>from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
<commit_msg>Add test to Path compare.<commit_after>
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
def test_cmp():
assert Path("/a/b") < Path("/a/c")
assert Path("/a/c") > Path("/a/b")
assert Path("/a/2") < Path("/b/1")
|
from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
Add test to Path compare.from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
def test_cmp():
assert Path("/a/b") < Path("/a/c")
assert Path("/a/c") > Path("/a/b")
assert Path("/a/2") < Path("/b/1")
|
<commit_before>from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
<commit_msg>Add test to Path compare.<commit_after>from farmfs.fs import normpath as _normalize
from farmfs.fs import userPath2Path as up2p
from farmfs.fs import Path
def test_normalize_abs():
assert _normalize("/") == "/"
assert _normalize("/a") == "/a"
assert _normalize("/a/") == "/a"
assert _normalize("/a/b") == "/a/b"
assert _normalize("/a/b/") == "/a/b"
assert _normalize("/a//b") == "/a/b"
assert _normalize("/a//b//") == "/a/b"
def test_normalize_relative():
assert _normalize("a") == "a"
assert _normalize("a/") == "a"
assert _normalize("a/b") == "a/b"
assert _normalize("a/b/") == "a/b"
assert _normalize("a//b") == "a/b"
assert _normalize("a//b//") == "a/b"
def test_userPath2Path():
assert up2p("c", Path("/a/b")) == Path("/a/b/c")
assert up2p("/c", Path("/a/b")) == Path("/c")
def test_cmp():
assert Path("/a/b") < Path("/a/c")
assert Path("/a/c") > Path("/a/b")
assert Path("/a/2") < Path("/b/1")
|
c41d0a9f03e66fdc20fb093aaad87cdd6f60461e
|
studies/helpers.py
|
studies/helpers.py
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
# TODO: celery taskify
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.celery import app
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
@app.task
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
Add decorator to send_mail function to celery taskify.
|
Add decorator to send_mail function to celery taskify.
|
Python
|
apache-2.0
|
pattisdr/lookit-api,pattisdr/lookit-api,CenterForOpenScience/lookit-api,CenterForOpenScience/lookit-api,pattisdr/lookit-api,CenterForOpenScience/lookit-api
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
# TODO: celery taskify
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
Add decorator to send_mail function to celery taskify.
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.celery import app
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
@app.task
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
<commit_before>from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
# TODO: celery taskify
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
<commit_msg>Add decorator to send_mail function to celery taskify.<commit_after>
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.celery import app
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
@app.task
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
# TODO: celery taskify
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
Add decorator to send_mail function to celery taskify.from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.celery import app
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
@app.task
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
<commit_before>from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
# TODO: celery taskify
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
<commit_msg>Add decorator to send_mail function to celery taskify.<commit_after>from django.core.mail.message import EmailMultiAlternatives
from django.template.loader import get_template
from project.celery import app
from project.settings import EMAIL_FROM_ADDRESS, BASE_URL
@app.task
def send_mail(template_name, subject, to_addresses, cc=None, bcc=None, from_email=None, **context):
"""
Helper for sending templated email
:param str template_name: Name of the template to send. There should exist a txt and html version
:param str subject: Subject line of the email
:param str from_address: From address for email
:param list to_addresses: List of addresses to email. If str is provided, wrapped in list
:param list cc: List of addresses to carbon copy
:param list bcc: List of addresses to blind carbon copy
:param str custom_message Custom email message - for use instead of a template
:kwargs: Context vars for the email template
"""
context['base_url'] = BASE_URL
text_content = get_template('emails/{}.txt'.format(template_name)).render(context)
html_content = get_template('emails/{}.html'.format(template_name)).render(context)
if not isinstance(to_addresses, list):
to_addresses = [to_addresses]
from_address = from_email or EMAIL_FROM_ADDRESS
email = EmailMultiAlternatives(subject, text_content, from_address, to_addresses, cc=cc, bcc=bcc)
email.attach_alternative(html_content, 'text/html')
email.send()
|
49a1548399fa822515920d910ec6ea6a6c813bca
|
threadpool.py
|
threadpool.py
|
from __future__ import with_statement
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
@threado.stream
def run(inner, self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
result = yield channel
inner.finish(result)
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
from __future__ import with_statement
import sys
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
def run(self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
return channel
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
|
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
|
Python
|
mit
|
abusesa/idiokit
|
from __future__ import with_statement
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
@threado.stream
def run(inner, self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
result = yield channel
inner.finish(result)
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.
|
from __future__ import with_statement
import sys
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
def run(self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
return channel
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
<commit_before>from __future__ import with_statement
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
@threado.stream
def run(inner, self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
result = yield channel
inner.finish(result)
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
<commit_msg>Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.<commit_after>
|
from __future__ import with_statement
import sys
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
def run(self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
return channel
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
from __future__ import with_statement
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
@threado.stream
def run(inner, self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
result = yield channel
inner.finish(result)
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.from __future__ import with_statement
import sys
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
def run(self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
return channel
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
<commit_before>from __future__ import with_statement
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
@threado.stream
def run(inner, self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
result = yield channel
inner.finish(result)
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
<commit_msg>Fix a weird situation when a function run in a thread raises StopIteration, which was erroneusly interpreted that the thread returned None.<commit_after>from __future__ import with_statement
import sys
import threado
import threading
import Queue
class ThreadPool(object):
def __init__(self, idle_time=5.0):
self.lock = threading.Lock()
self.threads = list()
self.idle_time = idle_time
def run(self, func, *args, **keys):
with self.lock:
if self.threads:
thread, queue = self.threads.pop()
else:
queue = Queue.Queue()
thread = threading.Thread(target=self._thread, args=(queue,))
thread.setDaemon(True)
channel = threado.Channel()
queue.put((channel, func, args, keys))
if not thread.isAlive():
thread.start()
return channel
def _thread(self, queue):
item = threading.currentThread(), queue
while True:
try:
task = queue.get(True, self.idle_time)
except Queue.Empty:
with self.lock:
if item not in self.threads:
continue
self.threads.remove(item)
return
if task is None:
return
channel, func, args, keys = task
try:
result = func(*args, **keys)
except:
self.threads.append(item)
channel.rethrow()
else:
self.threads.append(item)
channel.finish(result)
thread_pool = ThreadPool()
run = thread_pool.run
|
b252592eb40263994317d88ced43ddc4669a4975
|
tests/run_tests.py
|
tests/run_tests.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
main(args)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
return unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
result = main(args)
if not result.wasSuccessful():
return_code = 1
else:
return_code = 0
sys.exit(return_code)
|
Fix return code when running unittests.
|
Fix return code when running unittests.
|
Python
|
bsd-3-clause
|
AndiDog/sqlparse,actsasgeek/sqlparse,zhongdai/sqlparse,tailhook/sqlparse,AndiDog/sqlparse,Yelp/sqlparse,Yelp/sqlparse,benekastah/sqlparse,actsasgeek/sqlparse,tailhook/sqlparse,Yelp/sqlparse,MariaPet/sqlparse,payne/sqlparse,adamgreenhall/sqlparse,andialbrecht/sqlparse,tailhook/sqlparse,adamgreenhall/sqlparse,payne/sqlparse,adamgreenhall/sqlparse,payne/sqlparse,MariaPet/sqlparse,zhongdai/sqlparse,vmuriart/sqlparse,AndiDog/sqlparse,zhongdai/sqlparse,actsasgeek/sqlparse
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
main(args)
Fix return code when running unittests.
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
return unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
result = main(args)
if not result.wasSuccessful():
return_code = 1
else:
return_code = 0
sys.exit(return_code)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
main(args)
<commit_msg>Fix return code when running unittests.<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
return unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
result = main(args)
if not result.wasSuccessful():
return_code = 1
else:
return_code = 0
sys.exit(return_code)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
main(args)
Fix return code when running unittests.#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
return unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
result = main(args)
if not result.wasSuccessful():
return_code = 1
else:
return_code = 0
sys.exit(return_code)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
main(args)
<commit_msg>Fix return code when running unittests.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
"""Test runner for sqlparse."""
import optparse
import os
import sys
import unittest
test_mod = os.path.abspath(os.path.join(os.path.dirname(__file__), '../'))
if test_mod not in sys.path:
sys.path.insert(1, test_mod)
parser = optparse.OptionParser()
parser.add_option('-P', '--profile',
help='Create hotshot profile.',
action='store_true', default=False)
def main(args):
"""Create a TestSuite and run it."""
loader = unittest.TestLoader()
suite = unittest.TestSuite()
fnames = [os.path.split(f)[-1] for f in args]
for fname in os.listdir(os.path.dirname(__file__)):
if (not fname.startswith('test_') or not fname.endswith('.py')
or (fnames and fname not in fnames)):
continue
modname = os.path.splitext(fname)[0]
mod = __import__(os.path.splitext(fname)[0])
suite.addTests(loader.loadTestsFromModule(mod))
return unittest.TextTestRunner(verbosity=2).run(suite)
if __name__ == '__main__':
opts, args = parser.parse_args()
if opts.profile:
import hotshot
prof = hotshot.Profile("sqlparse.prof")
prof.runcall(main, args)
prof.close()
else:
result = main(args)
if not result.wasSuccessful():
return_code = 1
else:
return_code = 0
sys.exit(return_code)
|
12ec1cf9084789b9e2022eb0d1d55b553db06cb5
|
tests/test_util.py
|
tests/test_util.py
|
import util
from nose.tools import assert_equal
class TestPick():
def check(self, filenames, expected, k, randomized):
result = util.pick(filenames, k, randomized)
assert_equal(result, expected)
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
self.check(filenames, expected, k=None, randomized=False)
|
import util
from nose.tools import assert_equal, assert_true, raises
class TestPick():
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
result = util.pick(filenames, randomized=False)
assert_equal(result, expected)
def test_k_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt']
result = util.pick(filenames, k=3, randomized=False)
assert_equal(result, expected)
def test_all_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames)
assert_equal(sorted(filenames), sorted(result))
def test_k_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames, k=3)
for r in result:
assert_true(r in filenames)
@raises(ValueError)
def test_negative_k(self):
util.pick([], k=-2)
|
Fix unit test for util.py
|
Fix unit test for util.py
|
Python
|
mit
|
kemskems/otdet
|
import util
from nose.tools import assert_equal
class TestPick():
def check(self, filenames, expected, k, randomized):
result = util.pick(filenames, k, randomized)
assert_equal(result, expected)
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
self.check(filenames, expected, k=None, randomized=False)
Fix unit test for util.py
|
import util
from nose.tools import assert_equal, assert_true, raises
class TestPick():
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
result = util.pick(filenames, randomized=False)
assert_equal(result, expected)
def test_k_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt']
result = util.pick(filenames, k=3, randomized=False)
assert_equal(result, expected)
def test_all_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames)
assert_equal(sorted(filenames), sorted(result))
def test_k_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames, k=3)
for r in result:
assert_true(r in filenames)
@raises(ValueError)
def test_negative_k(self):
util.pick([], k=-2)
|
<commit_before>import util
from nose.tools import assert_equal
class TestPick():
def check(self, filenames, expected, k, randomized):
result = util.pick(filenames, k, randomized)
assert_equal(result, expected)
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
self.check(filenames, expected, k=None, randomized=False)
<commit_msg>Fix unit test for util.py<commit_after>
|
import util
from nose.tools import assert_equal, assert_true, raises
class TestPick():
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
result = util.pick(filenames, randomized=False)
assert_equal(result, expected)
def test_k_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt']
result = util.pick(filenames, k=3, randomized=False)
assert_equal(result, expected)
def test_all_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames)
assert_equal(sorted(filenames), sorted(result))
def test_k_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames, k=3)
for r in result:
assert_true(r in filenames)
@raises(ValueError)
def test_negative_k(self):
util.pick([], k=-2)
|
import util
from nose.tools import assert_equal
class TestPick():
def check(self, filenames, expected, k, randomized):
result = util.pick(filenames, k, randomized)
assert_equal(result, expected)
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
self.check(filenames, expected, k=None, randomized=False)
Fix unit test for util.pyimport util
from nose.tools import assert_equal, assert_true, raises
class TestPick():
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
result = util.pick(filenames, randomized=False)
assert_equal(result, expected)
def test_k_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt']
result = util.pick(filenames, k=3, randomized=False)
assert_equal(result, expected)
def test_all_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames)
assert_equal(sorted(filenames), sorted(result))
def test_k_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames, k=3)
for r in result:
assert_true(r in filenames)
@raises(ValueError)
def test_negative_k(self):
util.pick([], k=-2)
|
<commit_before>import util
from nose.tools import assert_equal
class TestPick():
def check(self, filenames, expected, k, randomized):
result = util.pick(filenames, k, randomized)
assert_equal(result, expected)
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
self.check(filenames, expected, k=None, randomized=False)
<commit_msg>Fix unit test for util.py<commit_after>import util
from nose.tools import assert_equal, assert_true, raises
class TestPick():
def test_all_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt', 'c-3.txt', 'a-4.txt']
result = util.pick(filenames, randomized=False)
assert_equal(result, expected)
def test_k_sequential(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
expected = ['e-0.txt', 'd-1.txt', 'b-2.txt']
result = util.pick(filenames, k=3, randomized=False)
assert_equal(result, expected)
def test_all_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames)
assert_equal(sorted(filenames), sorted(result))
def test_k_random(self):
filenames = ['a-4.txt', 'b-2.txt', 'c-3.txt', 'd-1.txt', 'e-0.txt']
result = util.pick(filenames, k=3)
for r in result:
assert_true(r in filenames)
@raises(ValueError)
def test_negative_k(self):
util.pick([], k=-2)
|
048107baf565c26a21421ef8f703be5cf7591cee
|
src/ggrc/fulltext/__init__.py
|
src/ggrc/fulltext/__init__.py
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags, **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags="", **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
Change tags to kwarg in full text record
|
Change tags to kwarg in full text record
This change will allow us to set the tags value with kwargs that are set
in custom attribute values.
|
Python
|
apache-2.0
|
edofic/ggrc-core,AleksNeStu/ggrc-core,josthkko/ggrc-core,andrei-karalionak/ggrc-core,AleksNeStu/ggrc-core,andrei-karalionak/ggrc-core,AleksNeStu/ggrc-core,kr41/ggrc-core,andrei-karalionak/ggrc-core,VinnieJohns/ggrc-core,edofic/ggrc-core,VinnieJohns/ggrc-core,j0gurt/ggrc-core,plamut/ggrc-core,edofic/ggrc-core,j0gurt/ggrc-core,plamut/ggrc-core,plamut/ggrc-core,VinnieJohns/ggrc-core,selahssea/ggrc-core,selahssea/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,AleksNeStu/ggrc-core,josthkko/ggrc-core,plamut/ggrc-core,andrei-karalionak/ggrc-core,selahssea/ggrc-core,kr41/ggrc-core,selahssea/ggrc-core,edofic/ggrc-core,josthkko/ggrc-core,kr41/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,VinnieJohns/ggrc-core
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags, **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
Change tags to kwarg in full text record
This change will allow us to set the tags value with kwargs that are set
in custom attribute values.
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags="", **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
<commit_before># Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags, **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
<commit_msg>Change tags to kwarg in full text record
This change will allow us to set the tags value with kwargs that are set
in custom attribute values.<commit_after>
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags="", **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags, **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
Change tags to kwarg in full text record
This change will allow us to set the tags value with kwargs that are set
in custom attribute values.# Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags="", **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
<commit_before># Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags, **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
<commit_msg>Change tags to kwarg in full text record
This change will allow us to set the tags value with kwargs that are set
in custom attribute values.<commit_after># Copyright (C) 2016 Google Inc.
# Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file>
from ggrc.extensions import get_extension_instance
class Indexer(object):
def __init__(self, settings):
pass
def create_record(self, record):
raise NotImplementedError()
def update_record(self, record):
raise NotImplementedError()
def delete_record(self, key):
raise NotImplementedError()
def search(self, terms):
raise NotImplementedError()
class Record(object):
def __init__(self, key, type, context_id, tags="", **kwargs):
self.key = key
self.type = type
self.context_id = context_id
self.tags = tags
self.properties = kwargs
def resolve_default_text_indexer():
from ggrc import settings
db_scheme = settings.SQLALCHEMY_DATABASE_URI.split(':')[0].split('+')[0]
return 'ggrc.fulltext.{db_scheme}.Indexer'.format(db_scheme=db_scheme)
def get_indexer(indexer=[]):
return get_extension_instance(
'FULLTEXT_INDEXER', resolve_default_text_indexer)
|
f6bcc5af968c5c8b81fd6db10bad50b1fadcb5d1
|
discover_road_runner/meep_meep/tests/test_meep_meep.py
|
discover_road_runner/meep_meep/tests/test_meep_meep.py
|
import unittest
from django.test import SimpleTestCase
class BasicTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
import unittest
from django.test import SimpleTestCase
class CombinedTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
Rename test as it really is a full integration test.
|
Rename test as it really is a full integration test.
|
Python
|
mit
|
pzrq/discover-road-runner
|
import unittest
from django.test import SimpleTestCase
class BasicTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
Rename test as it really is a full integration test.
|
import unittest
from django.test import SimpleTestCase
class CombinedTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
<commit_before>import unittest
from django.test import SimpleTestCase
class BasicTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
<commit_msg>Rename test as it really is a full integration test.<commit_after>
|
import unittest
from django.test import SimpleTestCase
class CombinedTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
import unittest
from django.test import SimpleTestCase
class BasicTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
Rename test as it really is a full integration test.import unittest
from django.test import SimpleTestCase
class CombinedTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
<commit_before>import unittest
from django.test import SimpleTestCase
class BasicTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
<commit_msg>Rename test as it really is a full integration test.<commit_after>import unittest
from django.test import SimpleTestCase
class CombinedTest(SimpleTestCase):
def test_success(self):
self.assertEqual(1 + 1, 2)
def test_error(self):
raise ValueError
def test_fail(self):
self.fail()
def test_skip(self):
self.skipTest('Fix me later?')
@unittest.expectedFailure
def test_decorator_expected_fail(self):
self.fail()
@unittest.skip
def test_decorator_skip(self):
print('Never get here!')
|
ff9a4f89d81c2059f51f7346700ce16972f04e36
|
django_basic_tinymce_flatpages/admin.py
|
django_basic_tinymce_flatpages/admin.py
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
default={'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
{'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
Fix getattr() takes no keyword arguments
|
Fix getattr() takes no keyword arguments
|
Python
|
bsd-3-clause
|
ad-m/django-basic-tinymce-flatpages
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
default={'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
Fix getattr() takes no keyword arguments
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
{'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
<commit_before>from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
default={'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
<commit_msg>Fix getattr() takes no keyword arguments<commit_after>
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
{'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
default={'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
Fix getattr() takes no keyword argumentsfrom django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
{'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
<commit_before>from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
default={'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
<commit_msg>Fix getattr() takes no keyword arguments<commit_after>from django.conf import settings
from django.contrib import admin
from django.contrib.flatpages.admin import FlatpageForm, FlatPageAdmin
from django.contrib.flatpages.models import FlatPage
from django.utils.module_loading import import_string
FLATPAGE_WIDGET = getattr(settings, 'FLATPAGE_WIDGET', 'tinymce.widgets.TinyMCE')
FLATPAGE_WIDGET_KWARGS = getattr(settings, 'FLATPAGE_WIDGET_KWARGS',
{'attrs': {'cols': 100, 'rows': 15}})
class PageForm(FlatpageForm):
class Meta:
model = FlatPage
widgets = {
'content': import_string(FLATPAGE_WIDGET)(**FLATPAGE_WIDGET_KWARGS),
}
class PageAdmin(FlatPageAdmin):
"""
Page Admin
"""
form = PageForm
admin.site.unregister(FlatPage)
admin.site.register(FlatPage, PageAdmin)
|
0bf00b40e84a5c5fbcdbeb7b81911998e3f1081a
|
src/idea/tests/smoke_tests.py
|
src/idea/tests/smoke_tests.py
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="test1@example.com", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "test1@example.com"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"test1@example.com"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="test1@example.com")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
Use fixtures for smoke tests
|
Use fixtures for smoke tests
|
Python
|
cc0-1.0
|
cmc333333/idea-box,m3brown/idea-box,18F/idea-box,cmc333333/idea-box,CapeSepias/idea-box,geomapdev/idea-box,CapeSepias/idea-box,cmc333333/idea-box,18F/idea-box,geomapdev/idea-box,geomapdev/idea-box,18F/idea-box,CapeSepias/idea-box,m3brown/idea-box
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="test1@example.com", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "test1@example.com"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"test1@example.com"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
Use fixtures for smoke tests
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="test1@example.com")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
<commit_before>import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="test1@example.com", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "test1@example.com"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"test1@example.com"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
<commit_msg>Use fixtures for smoke tests<commit_after>
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="test1@example.com")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="test1@example.com", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "test1@example.com"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"test1@example.com"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
Use fixtures for smoke testsimport os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="test1@example.com")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
<commit_before>import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state']
@fixture
def user(self):
try:
from collab.django_factories import UserF
return UserF(username="test1@example.com", person__title='')
except ImportError:
from django.contrib.auth.models import User
user = User()
user.username = "test1@example.com"
user.first_name = 'first'
user.last_name = 'last'
user.email = '"test1@example.com"'
user.password = 'pbkdf2_sha256$10000$ggAKkiHobFL8$xQzwPeHNX1vWr9uNmZ/gKbd17uLGZVM8QNcgmaIEAUs='
user.is_staff = False
user.is_active = True
user.is_superuser = False
user.last_login = timezone.now()
user.date_joined = timezone.now()
user.save()
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
<commit_msg>Use fixtures for smoke tests<commit_after>import os
from django.utils import timezone
from django_webtest import WebTest
from exam.decorators import fixture
from exam.cases import Exam
from django.core.urlresolvers import reverse
from django.contrib.auth.models import User
class SmokeTest(Exam, WebTest):
csrf_checks = False
fixtures = ['state', 'core-test-fixtures']
@fixture
def user(self):
user = User.objects.get(username="test1@example.com")
return user
def get(self, url):
return self.app.get(url, user=self.user)
def test_idea_home(self):
page = self.get(reverse('idea:idea_list'))
self.assertEquals(200, page.status_code)
|
b77a3f47876d824d2e0f1c009a6d580fc5d41ec6
|
accelerator/migrations/0019_add_deferred_user_role.py
|
accelerator/migrations/0019_add_deferred_user_role.py
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
user_role = UserRole.objects.create(
name=DEFERRED_MENTOR, sort_order='17')
for program in Program.objects.all():
name = "{} {} Deferred Mentor".format(
program.start_date.year,
program.program_family.name)
return ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
defaults={'name': name})
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
if UserRole.objects.filter(name=DEFERRED_MENTOR).exists():
user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0]
else:
user_role = UserRole.objects.create(name=DEFERRED_MENTOR,
sort_order=17)
for program in Program.objects.all():
if not ProgramRole.objects.filter(user_role=user_role,
program=program).exists():
name = "{} {} ({}-{})".format(
(program.end_date.year if program.end_date else ""),
DEFERRED_MENTOR,
program.program_family.url_slug.upper(),
program.pk)
ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
name=name)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
Make changes to the migration file
|
[AC-7594] Make changes to the migration file
|
Python
|
mit
|
masschallenge/django-accelerator,masschallenge/django-accelerator
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
user_role = UserRole.objects.create(
name=DEFERRED_MENTOR, sort_order='17')
for program in Program.objects.all():
name = "{} {} Deferred Mentor".format(
program.start_date.year,
program.program_family.name)
return ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
defaults={'name': name})
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
[AC-7594] Make changes to the migration file
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
if UserRole.objects.filter(name=DEFERRED_MENTOR).exists():
user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0]
else:
user_role = UserRole.objects.create(name=DEFERRED_MENTOR,
sort_order=17)
for program in Program.objects.all():
if not ProgramRole.objects.filter(user_role=user_role,
program=program).exists():
name = "{} {} ({}-{})".format(
(program.end_date.year if program.end_date else ""),
DEFERRED_MENTOR,
program.program_family.url_slug.upper(),
program.pk)
ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
name=name)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
<commit_before># Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
user_role = UserRole.objects.create(
name=DEFERRED_MENTOR, sort_order='17')
for program in Program.objects.all():
name = "{} {} Deferred Mentor".format(
program.start_date.year,
program.program_family.name)
return ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
defaults={'name': name})
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
<commit_msg>[AC-7594] Make changes to the migration file<commit_after>
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
if UserRole.objects.filter(name=DEFERRED_MENTOR).exists():
user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0]
else:
user_role = UserRole.objects.create(name=DEFERRED_MENTOR,
sort_order=17)
for program in Program.objects.all():
if not ProgramRole.objects.filter(user_role=user_role,
program=program).exists():
name = "{} {} ({}-{})".format(
(program.end_date.year if program.end_date else ""),
DEFERRED_MENTOR,
program.program_family.url_slug.upper(),
program.pk)
ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
name=name)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
user_role = UserRole.objects.create(
name=DEFERRED_MENTOR, sort_order='17')
for program in Program.objects.all():
name = "{} {} Deferred Mentor".format(
program.start_date.year,
program.program_family.name)
return ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
defaults={'name': name})
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
[AC-7594] Make changes to the migration file# Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
if UserRole.objects.filter(name=DEFERRED_MENTOR).exists():
user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0]
else:
user_role = UserRole.objects.create(name=DEFERRED_MENTOR,
sort_order=17)
for program in Program.objects.all():
if not ProgramRole.objects.filter(user_role=user_role,
program=program).exists():
name = "{} {} ({}-{})".format(
(program.end_date.year if program.end_date else ""),
DEFERRED_MENTOR,
program.program_family.url_slug.upper(),
program.pk)
ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
name=name)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
<commit_before># Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
user_role = UserRole.objects.create(
name=DEFERRED_MENTOR, sort_order='17')
for program in Program.objects.all():
name = "{} {} Deferred Mentor".format(
program.start_date.year,
program.program_family.name)
return ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
defaults={'name': name})
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
<commit_msg>[AC-7594] Make changes to the migration file<commit_after># Generated by Django 2.2.10 on 2020-04-09 21:24
from django.db import migrations
def add_deferred_user_role(apps, schema_editor):
DEFERRED_MENTOR = 'Deferred Mentor'
UserRole = apps.get_model('accelerator', 'UserRole')
Program = apps.get_model('accelerator', 'Program')
ProgramRole = apps.get_model('accelerator', 'ProgramRole')
if UserRole.objects.filter(name=DEFERRED_MENTOR).exists():
user_role = UserRole.objects.filter(user=DEFERRED_MENTOR)[0]
else:
user_role = UserRole.objects.create(name=DEFERRED_MENTOR,
sort_order=17)
for program in Program.objects.all():
if not ProgramRole.objects.filter(user_role=user_role,
program=program).exists():
name = "{} {} ({}-{})".format(
(program.end_date.year if program.end_date else ""),
DEFERRED_MENTOR,
program.program_family.url_slug.upper(),
program.pk)
ProgramRole.objects.get_or_create(
program=program,
user_role=user_role,
name=name)
class Migration(migrations.Migration):
dependencies = [
('accelerator', '0018_make_location_nonrequired'),
]
operations = [
migrations.RunPython(add_deferred_user_role,
migrations.RunPython.noop)
]
|
4574d25ade5c18d6c15ac6d427f4fbd4cb2f0f04
|
braid/info.py
|
braid/info.py
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
('rhel', '/etc/redhat-release'),
('debian', '/etc/debian_version'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
Add debian-squeeze support for os detection.
|
Add debian-squeeze support for os detection.
|
Python
|
mit
|
alex/braid,alex/braid
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
Add debian-squeeze support for os detection.
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
('rhel', '/etc/redhat-release'),
('debian', '/etc/debian_version'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
<commit_before>from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
<commit_msg>Add debian-squeeze support for os detection.<commit_after>
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
('rhel', '/etc/redhat-release'),
('debian', '/etc/debian_version'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
Add debian-squeeze support for os detection.from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
('rhel', '/etc/redhat-release'),
('debian', '/etc/debian_version'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
<commit_before>from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
<commit_msg>Add debian-squeeze support for os detection.<commit_after>from fabric.api import run, quiet
from braid import succeeds, cacheInEnvironment
@cacheInEnvironment
def distroName():
"""
Get the name of the distro.
"""
with quiet():
lsb = run('/usr/bin/lsb_release --id --short', warn_only=True)
if lsb.succeeded:
return lsb.lower()
distros = [
('centos', '/etc/centos-release'),
('fedora', '/etc/fedora-release'),
('rhel', '/etc/redhat-release'),
('debian', '/etc/debian_version'),
]
for distro, sentinel in distros:
if succeeds('/usr/bin/test -f {}'.format(sentinel)):
return distro
def distroFamily():
"""
Get the family of the distro.
@returns: C{'debian'} or C{'fedora'}
"""
families = {
'debian': ['debian', 'ubuntu'],
'fedora': ['fedora', 'centos', 'rhel'],
}
distro = distroName()
for family, members in families.iteritems():
if distro in members:
return family
return 'other'
@cacheInEnvironment
def arch():
"""
Get the architechture of the machine.
"""
return run('/bin/uname --machine')
@cacheInEnvironment
def isRoot():
"""
Check if the current user is root.
"""
return run('id -u') == '0'
__all__ = ['distroName', 'distroFamily', 'arch']
|
ee7ced467a7b87e71aa5a1df4c828e672d0b9870
|
Utils/py/BallDetection/RegressionNetwork/evaluate.py
|
Utils/py/BallDetection/RegressionNetwork/evaluate.py
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
print("loss: {} precision: {}".format(result[0], result[1]))
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
for idx in range(0, len(result)):
print(model.metrics_names[idx] + ":", result[idx])
|
Print all evaluation metrics associated with the model
|
Print all evaluation metrics associated with the model
|
Python
|
apache-2.0
|
BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH,BerlinUnited/NaoTH
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
print("loss: {} precision: {}".format(result[0], result[1]))
Print all evaluation metrics associated with the model
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
for idx in range(0, len(result)):
print(model.metrics_names[idx] + ":", result[idx])
|
<commit_before>#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
print("loss: {} precision: {}".format(result[0], result[1]))
<commit_msg>Print all evaluation metrics associated with the model<commit_after>
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
for idx in range(0, len(result)):
print(model.metrics_names[idx] + ":", result[idx])
|
#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
print("loss: {} precision: {}".format(result[0], result[1]))
Print all evaluation metrics associated with the model#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
for idx in range(0, len(result)):
print(model.metrics_names[idx] + ":", result[idx])
|
<commit_before>#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
print("loss: {} precision: {}".format(result[0], result[1]))
<commit_msg>Print all evaluation metrics associated with the model<commit_after>#!/usr/bin/env python3
import argparse
import pickle
import tensorflow.keras as keras
import numpy as np
import sys
import cv2
parser = argparse.ArgumentParser(description='Train the network given ')
parser.add_argument('-b', '--database-path', dest='imgdb_path',
help='Path to the image database containing test data.'
'Default is img.db in current folder.')
parser.add_argument('-m', '--model-path', dest='model_path',
help='Store the trained model using this path. Default is model.h5.')
args = parser.parse_args()
imgdb_path = "img.db"
model_path = "model.h5"
res = {"x": 16, "y": 16}
if args.model_path is not None:
model_path = args.model_path
if args.imgdb_path is not None:
imgdb_path = args.imgdb_path
with open(imgdb_path, "rb") as f:
mean = pickle.load(f)
print("mean=" + str(mean))
x = pickle.load(f)
y = pickle.load(f)
model = keras.models.load_model(model_path)
print(model.summary())
x = np.array(x)
y = np.array(y)
result = model.evaluate(x,y)
print("Evaluation result")
print("=================")
for idx in range(0, len(result)):
print(model.metrics_names[idx] + ":", result[idx])
|
d25603818e6af0b99ee1a6add0a7e182037d7a12
|
tests/test_set_pref.py
|
tests/test_set_pref.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs() as recent_refs:
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs():
yvs.shared.update_recent_refs(['8/mat.5'])
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
nose.assert_equal(len(yvs.shared.get_recent_refs()), 0)
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
Verify that recent list is cleared when language is changed
|
Verify that recent list is cleared when language is changed
|
Python
|
mit
|
caleb531/youversion-suggest,caleb531/youversion-suggest
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs() as recent_refs:
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
Verify that recent list is cleared when language is changed
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs():
yvs.shared.update_recent_refs(['8/mat.5'])
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
nose.assert_equal(len(yvs.shared.get_recent_refs()), 0)
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs() as recent_refs:
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
<commit_msg>Verify that recent list is cleared when language is changed<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs():
yvs.shared.update_recent_refs(['8/mat.5'])
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
nose.assert_equal(len(yvs.shared.get_recent_refs()), 0)
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs() as recent_refs:
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
Verify that recent list is cleared when language is changed#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs():
yvs.shared.update_recent_refs(['8/mat.5'])
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
nose.assert_equal(len(yvs.shared.get_recent_refs()), 0)
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs() as recent_refs:
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
# Check if new values have been saved to preferences
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
<commit_msg>Verify that recent list is cleared when language is changed<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
import nose.tools as nose
import yv_suggest.set_pref as yvs
import context_managers as ctx
def test_set_language():
"""should set preferred language"""
with ctx.preserve_prefs() as prefs:
with ctx.preserve_recent_refs():
yvs.shared.update_recent_refs(['8/mat.5'])
languages = yvs.shared.get_languages()
for language in languages:
if language['id'] != prefs['language']:
new_language = language['id']
break
yvs.main('language:{}'.format(new_language))
prefs = yvs.shared.get_prefs()
bible = yvs.shared.get_bible_data(prefs['language'])
nose.assert_equal(prefs['language'], new_language)
nose.assert_equal(prefs['version'], bible['default_version'])
nose.assert_equal(len(yvs.shared.get_recent_refs()), 0)
def test_set_version():
"""should set preferred version"""
with ctx.preserve_prefs() as prefs:
bible = yvs.shared.get_bible_data(prefs['language'])
versions = bible['versions']
for version in versions:
if version['id'] != prefs['language']:
new_version = version['id']
break
yvs.main('version:{}'.format(new_version))
prefs = yvs.shared.get_prefs()
nose.assert_equal(prefs['version'], new_version)
|
88098475358aaee18d32a1ad2c4a4301672bca0e
|
account_move_fiscal_month/models/account_move_line.py
|
account_move_fiscal_month/models/account_move_line.py
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
store=True, copy=False)
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
)
|
Remove unneeded and inefficient "store=True"
|
[FIX] Remove unneeded and inefficient "store=True"
|
Python
|
agpl-3.0
|
OCA/account-financial-tools,OCA/account-financial-tools
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
store=True, copy=False)
[FIX] Remove unneeded and inefficient "store=True"
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
)
|
<commit_before># Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
store=True, copy=False)
<commit_msg>[FIX] Remove unneeded and inefficient "store=True"<commit_after>
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
)
|
# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
store=True, copy=False)
[FIX] Remove unneeded and inefficient "store=True"# Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
)
|
<commit_before># Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
store=True, copy=False)
<commit_msg>[FIX] Remove unneeded and inefficient "store=True"<commit_after># Copyright 2017 ACSONE SA/NV
# License AGPL-3.0 or later (http://www.gnu.org/licenses/agpl).
from odoo import fields, models
class AccountMoveLine(models.Model):
_inherit = 'account.move.line'
date_range_fm_id = fields.Many2one(
related='move_id.date_range_fm_id',
)
|
68a7fd8a444a8c568d716db11849f58ad7a9dee5
|
django_pesapal/views.py
|
django_pesapal/views.py
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
Add support for further processing of the payment while maintaining compatibility
|
Add support for further processing of the payment while maintaining compatibility
|
Python
|
bsd-3-clause
|
odero/django-pesapal,odero/django-pesapal
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
Add support for further processing of the payment while maintaining compatibility
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
<commit_before># Create your views here.
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
<commit_msg>Add support for further processing of the payment while maintaining compatibility<commit_after>
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
# Create your views here.
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
Add support for further processing of the payment while maintaining compatibility# Create your views here.
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
<commit_before># Create your views here.
from django.core.urlresolvers import reverse_lazy
from django.contrib.auth.decorators import login_required
from django.views.generic.base import RedirectView
from django.db.models.loading import get_model
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if transaction_id and merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=merchant_reference, pesapal_transaction=transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
<commit_msg>Add support for further processing of the payment while maintaining compatibility<commit_after># Create your views here.
from django.core.urlresolvers import reverse_lazy, reverse
from django.views.generic.base import RedirectView
from django.core.urlresolvers import NoReverseMatch
from .models import Transaction
import conf
class TransactionCompletedView(RedirectView):
permanent = False
url = None
def get(self, request, *args, **kwargs):
'''
After Pesapal processes the transaction this will save the transaction and then redirect
to whatever reidrect URL in your settings as `PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL`.
For further processing just create a `post_save` signal on the `Transaction` model.
'''
self.transaction_id = request.GET.get('pesapal_transaction_tracking_id', '')
self.merchant_reference = request.GET.get('pesapal_merchant_reference', '')
if self.transaction_id and self.merchant_reference:
transaction, created = Transaction.objects.get_or_create(merchant_reference=self.merchant_reference,
pesapal_transaction=self.transaction_id)
return super(TransactionCompletedView, self).get(request, *args, **kwargs)
def get_redirect_url(self, *args, **kwargs):
try:
url = reverse(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL)
except NoReverseMatch:
url = reverse_lazy(conf.PESAPAL_TRANSACTION_DEFAULT_REDIRECT_URL,
kwargs={'merchant_reference': self.merchant_reference})
return url
|
7c1345c09cffaef9710bed24f6ae4ed791c3ade0
|
setup_cuda_deps.py
|
setup_cuda_deps.py
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'chainer',
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
Add chainer to required packages
|
Add chainer to required packages
|
Python
|
mit
|
yanweifu/chainer,ytoyama/yans_chainer_hackathon,AlpacaDB/chainer,jfsantos/chainer,jnishi/chainer,cupy/cupy,woodshop/chainer,okuta/chainer,woodshop/complex-chainer,niboshi/chainer,ysekky/chainer,cupy/cupy,hvy/chainer,ktnyt/chainer,hvy/chainer,wkentaro/chainer,benob/chainer,rezoo/chainer,kiyukuta/chainer,1986ks/chainer,Kaisuke5/chainer,chainer/chainer,minhpqn/chainer,truongdq/chainer,jnishi/chainer,tigerneil/chainer,pfnet/chainer,ikasumi/chainer,benob/chainer,wkentaro/chainer,anaruse/chainer,hvy/chainer,wkentaro/chainer,t-abe/chainer,cupy/cupy,okuta/chainer,delta2323/chainer,okuta/chainer,tkerola/chainer,keisuke-umezawa/chainer,keisuke-umezawa/chainer,muupan/chainer,laysakura/chainer,sinhrks/chainer,ktnyt/chainer,hvy/chainer,chainer/chainer,ktnyt/chainer,kashif/chainer,kuwa32/chainer,umitanuki/chainer,wkentaro/chainer,keisuke-umezawa/chainer,kikusu/chainer,niboshi/chainer,keisuke-umezawa/chainer,t-abe/chainer,kikusu/chainer,sinhrks/chainer,jnishi/chainer,muupan/chainer,niboshi/chainer,tscohen/chainer,elviswf/chainer,AlpacaDB/chainer,truongdq/chainer,cupy/cupy,chainer/chainer,masia02/chainer,wavelets/chainer,hidenori-t/chainer,ronekko/chainer,ktnyt/chainer,okuta/chainer,cemoody/chainer,chainer/chainer,bayerj/chainer,tereka114/chainer,jnishi/chainer,aonotas/chainer,sou81821/chainer,niboshi/chainer
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
Add chainer to required packages
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'chainer',
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
<commit_before>#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
<commit_msg>Add chainer to required packages<commit_after>
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'chainer',
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
Add chainer to required packages#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'chainer',
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
<commit_before>#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
<commit_msg>Add chainer to required packages<commit_after>#!/usr/bin/env python
from setuptools import setup
setup(
name='chainer-cuda-deps',
version='1.0.0',
description='Install dependent packages to use Chainer on CUDA',
author='Seiya Tokui',
author_email='tokui@preferred.jp',
url='http://chainer.org/',
packages=[],
install_requires=[
'chainer',
'pycuda>=2014.1',
'scikits.cuda>=0.5.0b1,!=0.042',
'Mako',
'six>=1.9.0',
],
)
|
b7106307baf97ba32cb29fe2a4bb9ed925c194ca
|
custom/onse/management/commands/update_onse_facility_cases.py
|
custom/onse/management/commands/update_onse_facility_cases.py
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(
print_notifications=True)
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(kwargs={
'print_notifications': True})
|
Fix passing keyword arg to task
|
Fix passing keyword arg to task
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(
print_notifications=True)
Fix passing keyword arg to task
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(kwargs={
'print_notifications': True})
|
<commit_before>from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(
print_notifications=True)
<commit_msg>Fix passing keyword arg to task<commit_after>
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(kwargs={
'print_notifications': True})
|
from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(
print_notifications=True)
Fix passing keyword arg to taskfrom django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(kwargs={
'print_notifications': True})
|
<commit_before>from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(
print_notifications=True)
<commit_msg>Fix passing keyword arg to task<commit_after>from django.core.management import BaseCommand
from custom.onse.tasks import update_facility_cases_from_dhis2_data_elements
class Command(BaseCommand):
help = ('Update facility_supervision cases with indicators collected '
'in DHIS2 over the last quarter.')
def handle(self, *args, **options):
update_facility_cases_from_dhis2_data_elements.apply(kwargs={
'print_notifications': True})
|
e5eff6f7f92b2946ca17e59c70b81df6f2e7a12d
|
opps/core/models/publisher.py
|
opps/core/models/publisher.py
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True)
date_update = models.DateTimeField(_(u"Date update"), auto_now=True)
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
kero = models.Manager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
def save(self, *args, **kwargs):
self.date_update = datetime.now()
super(Publisher, self).save(*args, **kwargs)
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
|
Remove date field (insert and update)
|
Remove date field (insert and update)
|
Python
|
mit
|
jeanmask/opps,jeanmask/opps,jeanmask/opps,YACOWS/opps,opps/opps,williamroot/opps,williamroot/opps,YACOWS/opps,opps/opps,YACOWS/opps,jeanmask/opps,opps/opps,williamroot/opps,opps/opps,YACOWS/opps,williamroot/opps
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True)
date_update = models.DateTimeField(_(u"Date update"), auto_now=True)
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
kero = models.Manager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
def save(self, *args, **kwargs):
self.date_update = datetime.now()
super(Publisher, self).save(*args, **kwargs)
Remove date field (insert and update)
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
|
<commit_before>#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True)
date_update = models.DateTimeField(_(u"Date update"), auto_now=True)
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
kero = models.Manager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
def save(self, *args, **kwargs):
self.date_update = datetime.now()
super(Publisher, self).save(*args, **kwargs)
<commit_msg>Remove date field (insert and update)<commit_after>
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
|
#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True)
date_update = models.DateTimeField(_(u"Date update"), auto_now=True)
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
kero = models.Manager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
def save(self, *args, **kwargs):
self.date_update = datetime.now()
super(Publisher, self).save(*args, **kwargs)
Remove date field (insert and update)#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
|
<commit_before>#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_insert = models.DateTimeField(_(u"Date insert"), auto_now_add=True)
date_update = models.DateTimeField(_(u"Date update"), auto_now=True)
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
kero = models.Manager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
def save(self, *args, **kwargs):
self.date_update = datetime.now()
super(Publisher, self).save(*args, **kwargs)
<commit_msg>Remove date field (insert and update)<commit_after>#!/usr/bin/env python
from django.db import models
from django.utils.translation import ugettext_lazy as _
from datetime import datetime
class PublisherMnager(models.Manager):
def all_published(self):
return super(PublisherMnager, self).get_query_set().filter(
date_available__lte=datetime.now(), published=True)
class Publisher(models.Model):
date_available = models.DateTimeField(_(u"Date available"),
default=datetime.now, null=True)
published = models.BooleanField(_(u"Published"), default=False)
objects = PublisherMnager()
class Meta:
abstract = True
def is_published(self):
return self.published and \
self.date_available.replace(tzinfo=None) <= datetime.now()
|
1dd58c6717fb8c3c23bce8cecf205c04cc03a134
|
comrade/views/simple.py
|
comrade/views/simple.py
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import Context, loader
from django.conf import settings
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(
t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import RequestContext, loader
from django.conf import settings
from maintenancemode.http import HttpResponseTemporaryUnavailable
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(t.render(RequestContext(request)))
def maintenance_mode(request, template_name='503.html'):
t = loader.get_template(template_name)
return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
|
Use requestcontext in error views.
|
Use requestcontext in error views.
|
Python
|
mit
|
bueda/django-comrade
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import Context, loader
from django.conf import settings
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(
t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
Use requestcontext in error views.
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import RequestContext, loader
from django.conf import settings
from maintenancemode.http import HttpResponseTemporaryUnavailable
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(t.render(RequestContext(request)))
def maintenance_mode(request, template_name='503.html'):
t = loader.get_template(template_name)
return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
|
<commit_before>from django.http import HttpResponse, HttpResponseServerError
from django.template import Context, loader
from django.conf import settings
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(
t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
<commit_msg>Use requestcontext in error views.<commit_after>
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import RequestContext, loader
from django.conf import settings
from maintenancemode.http import HttpResponseTemporaryUnavailable
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(t.render(RequestContext(request)))
def maintenance_mode(request, template_name='503.html'):
t = loader.get_template(template_name)
return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
|
from django.http import HttpResponse, HttpResponseServerError
from django.template import Context, loader
from django.conf import settings
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(
t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
Use requestcontext in error views.from django.http import HttpResponse, HttpResponseServerError
from django.template import RequestContext, loader
from django.conf import settings
from maintenancemode.http import HttpResponseTemporaryUnavailable
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(t.render(RequestContext(request)))
def maintenance_mode(request, template_name='503.html'):
t = loader.get_template(template_name)
return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
|
<commit_before>from django.http import HttpResponse, HttpResponseServerError
from django.template import Context, loader
from django.conf import settings
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(
t.render(Context({ 'MEDIA_URL': settings.MEDIA_URL })))
<commit_msg>Use requestcontext in error views.<commit_after>from django.http import HttpResponse, HttpResponseServerError
from django.template import RequestContext, loader
from django.conf import settings
from maintenancemode.http import HttpResponseTemporaryUnavailable
import logging
logger = logging.getLogger('comrade.views.simple')
def status(request):
logger.info("Responding to status check")
return HttpResponse()
def server_error(request, template_name='500.html'):
t = loader.get_template(template_name)
return HttpResponseServerError(t.render(RequestContext(request)))
def maintenance_mode(request, template_name='503.html'):
t = loader.get_template(template_name)
return HttpResponseTemporaryUnavailable(t.render(RequestContext(request)))
|
0ba671698bf4e268ae3f17e11078a5eb669a174c
|
indico/modules/events/roles/__init__.py
|
indico/modules/events/roles/__init__.py
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
roles_section = 'organization' if event.type == 'conference' else 'advanced'
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
|
Move roles menu item into a submenu
|
Move roles menu item into a submenu
- 'organization' for conferences
- 'advanced' for other event types
|
Python
|
mit
|
mic4ael/indico,indico/indico,mic4ael/indico,pferreir/indico,DirkHoffmann/indico,OmeGak/indico,pferreir/indico,OmeGak/indico,DirkHoffmann/indico,indico/indico,mvidalgarcia/indico,ThiefMaster/indico,indico/indico,OmeGak/indico,ThiefMaster/indico,DirkHoffmann/indico,DirkHoffmann/indico,mvidalgarcia/indico,ThiefMaster/indico,pferreir/indico,mvidalgarcia/indico,mvidalgarcia/indico,ThiefMaster/indico,pferreir/indico,mic4ael/indico,indico/indico,OmeGak/indico,mic4ael/indico
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
Move roles menu item into a submenu
- 'organization' for conferences
- 'advanced' for other event types
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
roles_section = 'organization' if event.type == 'conference' else 'advanced'
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
|
<commit_before># This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
<commit_msg>Move roles menu item into a submenu
- 'organization' for conferences
- 'advanced' for other event types<commit_after>
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
roles_section = 'organization' if event.type == 'conference' else 'advanced'
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
|
# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
Move roles menu item into a submenu
- 'organization' for conferences
- 'advanced' for other event types# This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
roles_section = 'organization' if event.type == 'conference' else 'advanced'
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
|
<commit_before># This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), 65, icon='medal')
<commit_msg>Move roles menu item into a submenu
- 'organization' for conferences
- 'advanced' for other event types<commit_after># This file is part of Indico.
# Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN).
#
# Indico is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 3 of the
# License, or (at your option) any later version.
#
# Indico is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Indico; if not, see <http://www.gnu.org/licenses/>.
from __future__ import unicode_literals
from flask import session
from indico.core import signals
from indico.core.logger import Logger
from indico.util.i18n import _
from indico.web.flask.util import url_for
from indico.web.menu import SideMenuItem
logger = Logger.get('events.roles')
@signals.menu.items.connect_via('event-management-sidemenu')
def _sidemenu_items(sender, event, **kwargs):
if event.can_manage(session.user):
roles_section = 'organization' if event.type == 'conference' else 'advanced'
return SideMenuItem('roles', _('Roles'), url_for('event_roles.manage', event), section=roles_section)
|
8ab4dd1012bedad286e19543f75e5a86d72a0402
|
biothings/web/api/es/handlers/biothing_handler.py
|
biothings/web/api/es/handlers/biothing_handler.py
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Inject bid(s) in esqb_kwargs.
Set GA tracking object.
'''
if self.request.method == 'GET':
options.esqb['bid'] = self.path_args[0] # TODO
elif self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish()
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Set GA tracking object.
'''
if self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_query_hook(self, options, query):
'''
Request _version field.
'''
options.es.version = True
return super().pre_query_hook(options, query)
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish() # TODO
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
Set _version requirement for annotation
|
Set _version requirement for annotation
|
Python
|
apache-2.0
|
biothings/biothings.api,biothings/biothings.api
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Inject bid(s) in esqb_kwargs.
Set GA tracking object.
'''
if self.request.method == 'GET':
options.esqb['bid'] = self.path_args[0] # TODO
elif self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish()
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
Set _version requirement for annotation
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Set GA tracking object.
'''
if self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_query_hook(self, options, query):
'''
Request _version field.
'''
options.es.version = True
return super().pre_query_hook(options, query)
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish() # TODO
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
<commit_before>
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Inject bid(s) in esqb_kwargs.
Set GA tracking object.
'''
if self.request.method == 'GET':
options.esqb['bid'] = self.path_args[0] # TODO
elif self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish()
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
<commit_msg>Set _version requirement for annotation<commit_after>
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Set GA tracking object.
'''
if self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_query_hook(self, options, query):
'''
Request _version field.
'''
options.es.version = True
return super().pre_query_hook(options, query)
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish() # TODO
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Inject bid(s) in esqb_kwargs.
Set GA tracking object.
'''
if self.request.method == 'GET':
options.esqb['bid'] = self.path_args[0] # TODO
elif self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish()
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
Set _version requirement for annotation
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Set GA tracking object.
'''
if self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_query_hook(self, options, query):
'''
Request _version field.
'''
options.es.version = True
return super().pre_query_hook(options, query)
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish() # TODO
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
<commit_before>
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Inject bid(s) in esqb_kwargs.
Set GA tracking object.
'''
if self.request.method == 'GET':
options.esqb['bid'] = self.path_args[0] # TODO
elif self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish()
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
<commit_msg>Set _version requirement for annotation<commit_after>
from tornado.web import Finish
from biothings.web.api.es.handlers.base_handler import ESRequestHandler
class BiothingHandler(ESRequestHandler):
'''
Handle requests to the annotation lookup endpoint.
'''
name = 'annotation'
def pre_query_builder_hook(self, options):
'''
Set GA tracking object.
'''
if self.request.method == 'POST':
self.ga_event_object({'qsize': len(options.esqb.ids)})
return options
def pre_query_hook(self, options, query):
'''
Request _version field.
'''
options.es.version = True
return super().pre_query_hook(options, query)
def pre_finish_hook(self, options, res):
'''
Return single result for GET.
Also does not return empty results.
'''
if isinstance(res, dict):
if not res.get('hits'):
self.send_error(404, reason=self.web_settings.ID_NOT_FOUND_TEMPLATE.format(bid=options.esqb.bid))
raise Finish() # TODO
res = res['hits'][0]
res.pop('_score')
elif isinstance(res, list):
for hit in res:
hit.pop('_score', None)
return res
|
369e70cc5d5e1c880bd61f055ffbe43d5aeab509
|
nodeconductor/structure/tests/unittests/test_models.py
|
nodeconductor/structure/tests/unittests/test_models.py
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
Break test flow into logical subsets
|
Break test flow into logical subsets [WAL-615]
|
Python
|
mit
|
opennode/nodeconductor,opennode/nodeconductor,opennode/nodeconductor
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
Break test flow into logical subsets [WAL-615]
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
<commit_before>from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
<commit_msg>Break test flow into logical subsets [WAL-615]<commit_after>
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
Break test flow into logical subsets [WAL-615]from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
<commit_before>from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
<commit_msg>Break test flow into logical subsets [WAL-615]<commit_after>from django.test import TestCase
from nodeconductor.structure.tests import factories
class ServiceProjectLinkTest(TestCase):
def setUp(self):
self.link = factories.TestServiceProjectLinkFactory()
def test_link_is_in_certification_erred_state_if_service_does_not_satisfy_project_certifications(self):
certification = factories.ServiceCertificationFactory()
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
self.link.project.certifications.add(certification)
self.assertEqual(self.link.CertificationState.ERRED, self.link.policy_compliant)
def test_link_is_in_certification_ok_state_if_project_certifications_is_a_subset_of_service_certifications(self):
certifications = factories.ServiceCertificationFactory.create_batch(2)
self.link.project.certifications.add(*certifications)
certifications.append(factories.ServiceCertificationFactory())
self.link.service.settings.certifications.add(*certifications)
self.assertEqual(self.link.CertificationState.OK, self.link.policy_compliant)
|
5a785f725d68733561a7e5e82c57655e25439ec8
|
indra/tests/test_grounding_resources.py
|
indra/tests/test_grounding_resources.py
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
|
Remove exceptional namespaces from test
|
Remove exceptional namespaces from test
|
Python
|
bsd-2-clause
|
johnbachman/indra,sorgerlab/indra,sorgerlab/belpy,sorgerlab/indra,bgyori/indra,sorgerlab/indra,bgyori/indra,bgyori/indra,johnbachman/indra,johnbachman/indra,sorgerlab/belpy,sorgerlab/belpy
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
Remove exceptional namespaces from test
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
|
<commit_before>import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
<commit_msg>Remove exceptional namespaces from test<commit_after>
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
|
import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
Remove exceptional namespaces from testimport os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
|
<commit_before>import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
# Namespaces that are not currently handled but still appear in statements
exceptions = ['CLO']
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_exceptional_unhandled():
"""Test that exceptional namespaces actually aren't handled.
This will catch if we make an update that makes an exceptional namespace
become a handled namespace. That way we can update the tests.
"""
actually_handled = []
for ns in exceptions:
if validate_ns(ns):
actually_handled.append(ns)
assert not actually_handled, actually_handled
<commit_msg>Remove exceptional namespaces from test<commit_after>import os
import csv
from indra.statements.validate import validate_db_refs, validate_ns
from indra.preassembler.grounding_mapper import default_grounding_map
from indra.preassembler.grounding_mapper import default_misgrounding_map
def test_misgrounding_map_entries():
bad_entries = []
for text, db_refs in default_misgrounding_map.items():
if not validate_db_refs(db_refs):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
def test_grounding_map_entries():
bad_entries = []
for text, db_refs in default_grounding_map.items():
if (not validate_db_refs(db_refs) and
not (set(exceptions) & db_refs.keys())):
bad_entries.append([text, db_refs])
assert not bad_entries, bad_entries
|
1e42bc1ef04ff3f52ce3f5db75d781be7d450a25
|
etl_framework/etl_class.py
|
etl_framework/etl_class.py
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.config
return getattr(config, key)
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif "config" in self.__dict__ and hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.__dict__["config"]
return getattr(config, key)
|
Make EtlClass attribute access more robust
|
Make EtlClass attribute access more robust
|
Python
|
mit
|
pantheon-systems/etl-framework
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.config
return getattr(config, key)
Make EtlClass attribute access more robust
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif "config" in self.__dict__ and hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.__dict__["config"]
return getattr(config, key)
|
<commit_before>"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.config
return getattr(config, key)
<commit_msg>Make EtlClass attribute access more robust<commit_after>
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif "config" in self.__dict__ and hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.__dict__["config"]
return getattr(config, key)
|
"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.config
return getattr(config, key)
Make EtlClass attribute access more robust"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif "config" in self.__dict__ and hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.__dict__["config"]
return getattr(config, key)
|
<commit_before>"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.config
return getattr(config, key)
<commit_msg>Make EtlClass attribute access more robust<commit_after>"""Base EtlClass that all EtlClasses should inherit"""
class EtlClass(object):
def __init__(self, config):
self.config = config
def __setattr__(self, key, value):
"""Set attribute on config if not in EtlClass object"""
if key == "config":
self.__dict__[key] = value
elif "config" in self.__dict__ and hasattr(self.config, key):
setattr(self.config, key, value)
else:
self.__dict__[key] = value
def __getattr__(self, key):
"""Get attribute on config if not in EtlClass object"""
# Get attribute if Config doesnt exist
# we don't need a special call to super here because getattr is only
# called when an attribute is NOT found in the instance's dictionary
config = self.__dict__["config"]
return getattr(config, key)
|
1f83113e748963cda9688d88a5d36dd7f9a54c1f
|
tests/app/test_cloudfoundry_config.py
|
tests/app/test_cloudfoundry_config.py
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def cloudfoundry_environ(os_environ, cloudfoundry_config):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ):
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def vcap_application(os_environ):
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
Move setting VCAP_SERVICES out of fixture
|
Move setting VCAP_SERVICES out of fixture
This was inconsistent with the source data for the fixture being
overidden in some of the tests. We only need to set it in the env
once, so it makes sense to just put the code there.
|
Python
|
mit
|
alphagov/notifications-api,alphagov/notifications-api
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def cloudfoundry_environ(os_environ, cloudfoundry_config):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ):
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
Move setting VCAP_SERVICES out of fixture
This was inconsistent with the source data for the fixture being
overidden in some of the tests. We only need to set it in the env
once, so it makes sense to just put the code there.
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def vcap_application(os_environ):
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
<commit_before>import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def cloudfoundry_environ(os_environ, cloudfoundry_config):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ):
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
<commit_msg>Move setting VCAP_SERVICES out of fixture
This was inconsistent with the source data for the fixture being
overidden in some of the tests. We only need to set it in the env
once, so it makes sense to just put the code there.<commit_after>
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def vcap_application(os_environ):
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def cloudfoundry_environ(os_environ, cloudfoundry_config):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ):
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
Move setting VCAP_SERVICES out of fixture
This was inconsistent with the source data for the fixture being
overidden in some of the tests. We only need to set it in the env
once, so it makes sense to just put the code there.import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def vcap_application(os_environ):
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
<commit_before>import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def cloudfoundry_environ(os_environ, cloudfoundry_config):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_environ):
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, cloudfoundry_environ):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
<commit_msg>Move setting VCAP_SERVICES out of fixture
This was inconsistent with the source data for the fixture being
overidden in some of the tests. We only need to set it in the env
once, so it makes sense to just put the code there.<commit_after>import json
import os
import pytest
from app.cloudfoundry_config import (
extract_cloudfoundry_config,
set_config_env_vars,
)
@pytest.fixture
def cloudfoundry_config():
return {
'postgres': [{
'credentials': {
'uri': 'postgres uri'
}
}],
'user-provided': []
}
@pytest.fixture
def vcap_application(os_environ):
os.environ['VCAP_APPLICATION'] = '{"space_name": "🚀🌌"}'
def test_extract_cloudfoundry_config_populates_other_vars(cloudfoundry_config, vcap_application):
os.environ['VCAP_SERVICES'] = json.dumps(cloudfoundry_config)
extract_cloudfoundry_config()
assert os.environ['SQLALCHEMY_DATABASE_URI'] == 'postgresql uri'
assert os.environ['NOTIFY_ENVIRONMENT'] == '🚀🌌'
assert os.environ['NOTIFY_LOG_PATH'] == '/home/vcap/logs/app.log'
def test_set_config_env_vars_ignores_unknown_configs(cloudfoundry_config, vcap_application):
cloudfoundry_config['foo'] = {'credentials': {'foo': 'foo'}}
cloudfoundry_config['user-provided'].append({
'name': 'bar', 'credentials': {'bar': 'bar'}
})
set_config_env_vars(cloudfoundry_config)
assert 'foo' not in os.environ
assert 'bar' not in os.environ
|
7cbe2351c2ad93def98005597a24e21d878ea492
|
flask_velox/mixins/http.py
|
flask_velox/mixins/http.py
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self, *args, **kwargs):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
Allow RedirectMixin to work within flask-admin
|
Allow RedirectMixin to work within flask-admin
|
Python
|
mit
|
thisissoon/Flask-Velox,thisissoon/Flask-Velox,jstacoder/Flask-Velox,jstacoder/Flask-Velox
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
Allow RedirectMixin to work within flask-admin
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self, *args, **kwargs):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
<commit_before># -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
<commit_msg>Allow RedirectMixin to work within flask-admin<commit_after>
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self, *args, **kwargs):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
Allow RedirectMixin to work within flask-admin# -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self, *args, **kwargs):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
<commit_before># -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
<commit_msg>Allow RedirectMixin to work within flask-admin<commit_after># -*- coding: utf-8 -*-
""" Module provides mixins for issuing HTTP Status codes using the
Flask ``View``.
"""
from flask import url_for
from flask.views import View
from werkzeug.utils import redirect
class RedirectMixin(View):
""" Raise a HTTP Redirect, by default a 302 HTTP Status Code will be used
however this can be overridden using the ``code`` attribute.
Example
-------
.. code-block:: python
:linenos:
from flask.ext.velox.mixins.http import RedirectMixin
class MyView(RedirectMixin):
rule = 'some.url.rule'
code = 301
Attributes
----------
rule : str
Flask URL Rule passed into ``url_for``
code : int, optional
Status code to raise, defaults to ``302``
"""
code = 302
def pre_dispatch(self, *args, **kwargs):
""" If you wish to run an arbitrary piece of code before the
redirect is dispatched you can override this method which is
called before dispatch.
"""
pass
def get_url(self):
""" Return a generated url from ``rule`` attribute.
Returns
-------
str
Generated url
"""
try:
rule = self.rule
except AttributeError:
raise NotImplementedError('``rule`` attr must be defined.')
return url_for(rule)
def dispatch_request(self, *args, **kwargs):
""" Dispatch the request, returning the redirect.func_closure
Returns
-------
werkzeug.wrappers.Response
Redirect response
"""
self.pre_dispatch()
return redirect(self.get_url(), code=getattr(self, 'code', 302))
|
c058ea00594efcdb39bb106ebe79dde39354e159
|
downloader.py
|
downloader.py
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "0",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "9",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
Use low quality audio (sounds good enough)
|
Use low quality audio (sounds good enough)
|
Python
|
mit
|
xrisk/webdl,xrisk/webdl,xrisk/webdl
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "0",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
Use low quality audio (sounds good enough)
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "9",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
<commit_before>import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "0",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
<commit_msg>Use low quality audio (sounds good enough)<commit_after>
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "9",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "0",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
Use low quality audio (sounds good enough)import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "9",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
<commit_before>import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "0",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
<commit_msg>Use low quality audio (sounds good enough)<commit_after>import sha
import os
def sha_hash(content):
return sha.new(content).hexdigest()
def download_audio(url):
from subprocess import call
retcode = call(["youtube-dl", "-x", "--id",
"--audio-quality", "9",
"--audio-format", "mp3",
"--exec",
"mv {} " + os.path.join('/app/mp3cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
def download_video(url):
from subprocess import call
retcode = call(["youtube-dl", "--format", "mp4",
"--exec",
"mv {} " + os.path.join('/app/mp4cache/', sha_hash(url)),
url])
if retcode == 0:
return sha_hash(url)
else:
raise Exception
|
bdda5e565981ac26a7e5e1ab8d1486eb91b09e4c
|
views/base.py
|
views/base.py
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path):
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path=None):
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
Use request.path if no path was passed to the default view
|
Use request.path if no path was passed to the default view
|
Python
|
bsd-3-clause
|
mjl/feincms,nickburlett/feincms,nickburlett/feincms,pjdelport/feincms,matthiask/django-content-editor,michaelkuty/feincms,matthiask/feincms2-content,mjl/feincms,matthiask/django-content-editor,matthiask/django-content-editor,mjl/feincms,hgrimelid/feincms,hgrimelid/feincms,joshuajonah/feincms,nickburlett/feincms,matthiask/django-content-editor,joshuajonah/feincms,joshuajonah/feincms,pjdelport/feincms,michaelkuty/feincms,matthiask/feincms2-content,feincms/feincms,pjdelport/feincms,feincms/feincms,michaelkuty/feincms,nickburlett/feincms,joshuajonah/feincms,feincms/feincms,matthiask/feincms2-content,michaelkuty/feincms,hgrimelid/feincms
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path):
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
Use request.path if no path was passed to the default view
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path=None):
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
<commit_before>from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path):
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
<commit_msg>Use request.path if no path was passed to the default view<commit_after>
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path=None):
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path):
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
Use request.path if no path was passed to the default viewfrom django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path=None):
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
<commit_before>from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path):
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
<commit_msg>Use request.path if no path was passed to the default view<commit_after>from django.http import HttpResponseRedirect
from django.shortcuts import render_to_response
from django.template import RequestContext
from feincms.models import Page
def handler(request, path=None):
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
if page.override_url:
return HttpResponseRedirect(page.override_url)
return render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request))
|
31bfe8fb498ea2e528da6463c9045b397992e028
|
python/caffe/test/test_draw.py
|
python/caffe/test/test_draw.py
|
import os
import unittest
from google import protobuf
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
protobuf.text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
|
import os
import unittest
from google.protobuf import text_format
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
if __name__ == "__main__":
unittest.main()
|
Add main() for draw_net unittest, fix import errors
|
Add main() for draw_net unittest, fix import errors
|
Python
|
apache-2.0
|
gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina,gnina/gnina
|
import os
import unittest
from google import protobuf
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
protobuf.text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
Add main() for draw_net unittest, fix import errors
|
import os
import unittest
from google.protobuf import text_format
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
if __name__ == "__main__":
unittest.main()
|
<commit_before>import os
import unittest
from google import protobuf
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
protobuf.text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
<commit_msg>Add main() for draw_net unittest, fix import errors<commit_after>
|
import os
import unittest
from google.protobuf import text_format
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
if __name__ == "__main__":
unittest.main()
|
import os
import unittest
from google import protobuf
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
protobuf.text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
Add main() for draw_net unittest, fix import errorsimport os
import unittest
from google.protobuf import text_format
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
if __name__ == "__main__":
unittest.main()
|
<commit_before>import os
import unittest
from google import protobuf
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
protobuf.text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
<commit_msg>Add main() for draw_net unittest, fix import errors<commit_after>import os
import unittest
from google.protobuf import text_format
import caffe.draw
from caffe.proto import caffe_pb2
def getFilenames():
"""Yields files in the source tree which are Net prototxts."""
result = []
root_dir = os.path.abspath(os.path.join(
os.path.dirname(__file__), '..', '..', '..'))
assert os.path.exists(root_dir)
for dirname in ('models', 'examples'):
dirname = os.path.join(root_dir, dirname)
assert os.path.exists(dirname)
for cwd, _, filenames in os.walk(dirname):
for filename in filenames:
filename = os.path.join(cwd, filename)
if filename.endswith('.prototxt') and 'solver' not in filename:
yield os.path.join(dirname, filename)
class TestDraw(unittest.TestCase):
def test_draw_net(self):
for filename in getFilenames():
net = caffe_pb2.NetParameter()
with open(filename) as infile:
text_format.Merge(infile.read(), net)
caffe.draw.draw_net(net, 'LR')
if __name__ == "__main__":
unittest.main()
|
b64bac9def90d53e76a3427f04f8664d513ae7ed
|
wallace/db.py
|
wallace/db.py
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
Fix assignment of database url
|
Fix assignment of database url
|
Python
|
mit
|
berkeley-cocosci/Wallace,Dallinger/Dallinger,berkeley-cocosci/Wallace,jcpeterson/Dallinger,suchow/Wallace,berkeley-cocosci/Wallace,Dallinger/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,jcpeterson/Dallinger,suchow/Wallace,suchow/Wallace,Dallinger/Dallinger,Dallinger/Dallinger,jcpeterson/Dallinger,Dallinger/Dallinger
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
Fix assignment of database url
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
<commit_before>"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
<commit_msg>Fix assignment of database url<commit_after>
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
Fix assignment of database url"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
<commit_before>"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
<commit_msg>Fix assignment of database url<commit_after>"""Create a connection to the database."""
from sqlalchemy import create_engine
from sqlalchemy.orm import sessionmaker, scoped_session
from sqlalchemy.ext.declarative import declarative_base
import os
db_url = os.environ.get("DATABASE_URL", "postgresql://postgres@localhost/wallace")
engine = create_engine(db_url)
Session = scoped_session(sessionmaker(autoflush=True, bind=engine))
Base = declarative_base()
Base.query = Session.query_property()
def init_db(drop_all=False):
"""Initialize the database, optionally dropping existing tables."""
if drop_all:
Base.metadata.drop_all(bind=engine)
Base.metadata.create_all(bind=engine)
return Session
|
e452663965f63ffb1d1e3ac04ae10dc26b3e2ecc
|
fabric/api.py
|
fabric/api.py
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env
from fabric.utils import abort, warn
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env, output
from fabric.utils import abort, warn
|
Allow access to state.output via API.
|
Allow access to state.output via API.
|
Python
|
bsd-2-clause
|
bitmonk/fabric,fernandezcuesta/fabric,opavader/fabric,xLegoz/fabric,getsentry/fabric,bitprophet/fabric,jaraco/fabric,MjAbuz/fabric,SamuelMarks/fabric,pashinin/fabric,tolbkni/fabric,pgroudas/fabric,bspink/fabric,likesxuqiang/fabric,hrubi/fabric,TarasRudnyk/fabric,sdelements/fabric,tekapo/fabric,simon-engledew/fabric,ericholscher/fabric,askulkarni2/fabric,ploxiln/fabric,kxxoling/fabric,haridsv/fabric,amaniak/fabric,kmonsoor/fabric,rodrigc/fabric,qinrong/fabric,cmattoon/fabric,cgvarela/fabric,raimon49/fabric,mathiasertl/fabric,akaariai/fabric,elijah513/fabric,rane-hs/fabric-py3,rbramwell/fabric,felix-d/fabric,itoed/fabric,StackStorm/fabric
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env
from fabric.utils import abort, warn
Allow access to state.output via API.
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env, output
from fabric.utils import abort, warn
|
<commit_before>"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env
from fabric.utils import abort, warn
<commit_msg>Allow access to state.output via API.<commit_after>
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env, output
from fabric.utils import abort, warn
|
"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env
from fabric.utils import abort, warn
Allow access to state.output via API."""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env, output
from fabric.utils import abort, warn
|
<commit_before>"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env
from fabric.utils import abort, warn
<commit_msg>Allow access to state.output via API.<commit_after>"""
Non-init module for doing convenient * imports from.
Necessary because if we did this in __init__, one would be unable to import
anything else inside the package -- like, say, the version number used in
setup.py -- without triggering loads of most of the code. Which doesn't work so
well when you're using setup.py to install e.g. paramiko!
"""
from fabric.context_managers import cd, hide, settings, show
from fabric.decorators import hosts, roles, runs_once
from fabric.operations import require, prompt, put, get, run, sudo, local
from fabric.state import env, output
from fabric.utils import abort, warn
|
d1e1f63062eff158b9bce8b9c3cbcaef1abed8ba
|
flask_gzip.py
|
flask_gzip.py
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers:
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if response.direct_passthrough:
return response
if (response.status_code not in xrange(200, 300) or
len(response.data) < self.minimum_size or
'Content-Encoding' in response.headers):
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level,
fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
Fix a runtime error when direct_passthrough is used.
|
Fix a runtime error when direct_passthrough is used.
|
Python
|
mit
|
wichitacode/flask-compress,wichitacode/flask-compress,libwilliam/flask-compress,saymedia/flask-compress,saymedia/flask-compress,libwilliam/flask-compress,libwilliam/flask-compress
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers:
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
Fix a runtime error when direct_passthrough is used.
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if response.direct_passthrough:
return response
if (response.status_code not in xrange(200, 300) or
len(response.data) < self.minimum_size or
'Content-Encoding' in response.headers):
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level,
fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
<commit_before>import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers:
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
<commit_msg>Fix a runtime error when direct_passthrough is used.<commit_after>
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if response.direct_passthrough:
return response
if (response.status_code not in xrange(200, 300) or
len(response.data) < self.minimum_size or
'Content-Encoding' in response.headers):
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level,
fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers:
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
Fix a runtime error when direct_passthrough is used.import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if response.direct_passthrough:
return response
if (response.status_code not in xrange(200, 300) or
len(response.data) < self.minimum_size or
'Content-Encoding' in response.headers):
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level,
fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
<commit_before>import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if (200 > response.status_code >= 300) or len(response.data) < self.minimum_size or 'Content-Encoding' in response.headers:
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level, fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
<commit_msg>Fix a runtime error when direct_passthrough is used.<commit_after>import gzip
import StringIO
from flask import request
class Gzip(object):
def __init__(self, app, compress_level=6, minimum_size=500):
self.app = app
self.compress_level = compress_level
self.minimum_size = minimum_size
self.app.after_request(self.after_request)
def after_request(self, response):
accept_encoding = request.headers.get('Accept-Encoding', '')
if 'gzip' not in accept_encoding.lower():
return response
if response.direct_passthrough:
return response
if (response.status_code not in xrange(200, 300) or
len(response.data) < self.minimum_size or
'Content-Encoding' in response.headers):
return response
gzip_buffer = StringIO.StringIO()
gzip_file = gzip.GzipFile(mode='wb', compresslevel=self.compress_level,
fileobj=gzip_buffer)
gzip_file.write(response.data)
gzip_file.close()
response.data = gzip_buffer.getvalue()
response.headers['Content-Encoding'] = 'gzip'
response.headers['Content-Length'] = len(response.data)
return response
|
23ad531d932b6c042c3bd0161b74a6088d02524f
|
myfedora/lib/app_globals.py
|
myfedora/lib/app_globals.py
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
from myfedora.streams import DataStreamer
self.datastreamer = DataStreamer()
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
#from myfedora.streams import DataStreamer
#self.datastreamer = DataStreamer()
FEED_CACHE = "/tmp/moksha-feeds"
from shove import Shove
from feedcache.cache import Cache
# is this not multi-process safe? or even thread safe?
self.feed_storage = Shove('file://' + FEED_CACHE)
self.feed_cache = Cache(self.feed_storage)
|
Add a feed_storage and feed_cache to our Globals object.
|
Add a feed_storage and feed_cache to our Globals object.
|
Python
|
agpl-3.0
|
fedora-infra/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages,Fale/fedora-packages,fedora-infra/fedora-packages,fedora-infra/fedora-packages,Fale/fedora-packages
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
from myfedora.streams import DataStreamer
self.datastreamer = DataStreamer()
Add a feed_storage and feed_cache to our Globals object.
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
#from myfedora.streams import DataStreamer
#self.datastreamer = DataStreamer()
FEED_CACHE = "/tmp/moksha-feeds"
from shove import Shove
from feedcache.cache import Cache
# is this not multi-process safe? or even thread safe?
self.feed_storage = Shove('file://' + FEED_CACHE)
self.feed_cache = Cache(self.feed_storage)
|
<commit_before>"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
from myfedora.streams import DataStreamer
self.datastreamer = DataStreamer()
<commit_msg>Add a feed_storage and feed_cache to our Globals object.<commit_after>
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
#from myfedora.streams import DataStreamer
#self.datastreamer = DataStreamer()
FEED_CACHE = "/tmp/moksha-feeds"
from shove import Shove
from feedcache.cache import Cache
# is this not multi-process safe? or even thread safe?
self.feed_storage = Shove('file://' + FEED_CACHE)
self.feed_cache = Cache(self.feed_storage)
|
"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
from myfedora.streams import DataStreamer
self.datastreamer = DataStreamer()
Add a feed_storage and feed_cache to our Globals object."""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
#from myfedora.streams import DataStreamer
#self.datastreamer = DataStreamer()
FEED_CACHE = "/tmp/moksha-feeds"
from shove import Shove
from feedcache.cache import Cache
# is this not multi-process safe? or even thread safe?
self.feed_storage = Shove('file://' + FEED_CACHE)
self.feed_cache = Cache(self.feed_storage)
|
<commit_before>"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
from myfedora.streams import DataStreamer
self.datastreamer = DataStreamer()
<commit_msg>Add a feed_storage and feed_cache to our Globals object.<commit_after>"""The application's Globals object"""
from app_factory import AppFactoryDict
class Globals(object):
"""Globals acts as a container for objects available throughout the
life of the application
"""
def __init__(self):
"""One instance of Globals is created during application
initialization and is available during requests via the 'g'
variable
"""
self.widgets = {'home': {}, 'canvas': {}, 'profile': {}, 'preview': {}, 'config':{}} # {viewtype: {name: Widget instance}}
self.resourceviews = AppFactoryDict() # {name: ResourceView instance}
self.apps = AppFactoryDict() # {name: App instance}
# Our comet data streamer, responsible for polling the data
# streams, and providing data to the widgets
#from myfedora.streams import DataStreamer
#self.datastreamer = DataStreamer()
FEED_CACHE = "/tmp/moksha-feeds"
from shove import Shove
from feedcache.cache import Cache
# is this not multi-process safe? or even thread safe?
self.feed_storage = Shove('file://' + FEED_CACHE)
self.feed_cache = Cache(self.feed_storage)
|
e9f3b6f9eb59ef7290498e8ceaf81c2bc66c8f59
|
ichnaea/gunicorn_config.py
|
ichnaea/gunicorn_config.py
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
# It should be 60 seconds, but gunicorn halves the configured value,
# see https://github.com/benoitc/gunicorn/issues/829
timeout = 120
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
timeout = 60
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
Update gunicorn timeout after gunicorn issue was answered.
|
Update gunicorn timeout after gunicorn issue was answered.
|
Python
|
apache-2.0
|
mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea,therewillbecode/ichnaea,mozilla/ichnaea
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
# It should be 60 seconds, but gunicorn halves the configured value,
# see https://github.com/benoitc/gunicorn/issues/829
timeout = 120
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
Update gunicorn timeout after gunicorn issue was answered.
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
timeout = 60
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
<commit_before># This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
# It should be 60 seconds, but gunicorn halves the configured value,
# see https://github.com/benoitc/gunicorn/issues/829
timeout = 120
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
<commit_msg>Update gunicorn timeout after gunicorn issue was answered.<commit_after>
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
timeout = 60
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
# It should be 60 seconds, but gunicorn halves the configured value,
# see https://github.com/benoitc/gunicorn/issues/829
timeout = 120
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
Update gunicorn timeout after gunicorn issue was answered.# This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
timeout = 60
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
<commit_before># This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
# It should be 60 seconds, but gunicorn halves the configured value,
# see https://github.com/benoitc/gunicorn/issues/829
timeout = 120
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
<commit_msg>Update gunicorn timeout after gunicorn issue was answered.<commit_after># This file contains gunicorn configuration setttings, as described at
# http://docs.gunicorn.org/en/latest/settings.html
# The file is loaded via the -c ichnaea.gunicorn_config command line option
# Be explicit about the worker class
worker_class = "sync"
# Set timeout to the same value as the default one from Amazon ELB (60 secs).
timeout = 60
# Recycle worker processes after 100k requests to prevent memory leaks
# from effecting us
max_requests = 100000
# Avoid too much output on the console
loglevel = "warning"
def post_worker_init(worker):
from random import randint
# Use 10% jitter, to prevent all workers from restarting at once,
# as they get an almost equal number of requests
jitter = randint(0, max_requests // 10)
worker.max_requests += jitter
# Actually initialize the application
worker.wsgi(None, None)
|
1c8f29d78d6409ba58df36d439f1ffd436c9dd10
|
gaphas/picklers.py
|
gaphas/picklers.py
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import new
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return new.instancemethod(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import types
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return types.MethodType(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
Fix no module 'new', replaced new.instancemethod with types.MethodType
|
Fix no module 'new', replaced new.instancemethod with types.MethodType
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
|
Python
|
lgpl-2.1
|
amolenaar/gaphas
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import new
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return new.instancemethod(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
Fix no module 'new', replaced new.instancemethod with types.MethodType
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import types
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return types.MethodType(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
<commit_before>"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import new
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return new.instancemethod(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
<commit_msg>Fix no module 'new', replaced new.instancemethod with types.MethodType
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import types
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return types.MethodType(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import new
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return new.instancemethod(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
Fix no module 'new', replaced new.instancemethod with types.MethodType
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import types
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return types.MethodType(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
<commit_before>"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import new
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return new.instancemethod(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(new.instancemethod, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
<commit_msg>Fix no module 'new', replaced new.instancemethod with types.MethodType
Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>"""
Some extra picklers needed to gracefully dump and load a canvas.
"""
from future import standard_library
standard_library.install_aliases()
import copyreg
# Allow instancemethod to be pickled:
import types
def construct_instancemethod(funcname, self, clazz):
func = getattr(clazz, funcname)
return types.MethodType(func, self, clazz)
def reduce_instancemethod(im):
return construct_instancemethod, (im.__func__.__name__, im.__self__, im.__self__.__class__)
copyreg.pickle(types.MethodType, reduce_instancemethod, construct_instancemethod)
# Allow cairo.Matrix to be pickled:
import cairo
def construct_cairo_matrix(*args):
return cairo.Matrix(*args)
def reduce_cairo_matrix(m):
return construct_cairo_matrix, tuple(m)
copyreg.pickle(cairo.Matrix, reduce_cairo_matrix, construct_cairo_matrix)
# vim:sw=4:et:ai
|
62f6e4a07ef63d53e1ff4230b7a40ad38168d103
|
python/testData/highlighting/assignmentTargets3K.py
|
python/testData/highlighting/assignmentTargets3K.py
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="assignment to keyword">True</error> = 2
<error descr="assignment to keyword">False</error> = 1
<error descr="assignment to keyword">None</error> = 0
<error descr="assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="Assignment to keyword">True</error> = 2
<error descr="Assignment to keyword">False</error> = 1
<error descr="Assignment to keyword">None</error> = 0
<error descr="Assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
Fix test broken by change in message capitalization
|
Fix test broken by change in message capitalization
GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda
|
Python
|
apache-2.0
|
allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community,allotria/intellij-community
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="assignment to keyword">True</error> = 2
<error descr="assignment to keyword">False</error> = 1
<error descr="assignment to keyword">None</error> = 0
<error descr="assignment to keyword">__debug__</error> = -1
() = []
[] = []
Fix test broken by change in message capitalization
GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="Assignment to keyword">True</error> = 2
<error descr="Assignment to keyword">False</error> = 1
<error descr="Assignment to keyword">None</error> = 0
<error descr="Assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
<commit_before><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="assignment to keyword">True</error> = 2
<error descr="assignment to keyword">False</error> = 1
<error descr="assignment to keyword">None</error> = 0
<error descr="assignment to keyword">__debug__</error> = -1
() = []
[] = []
<commit_msg>Fix test broken by change in message capitalization
GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<commit_after>
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="Assignment to keyword">True</error> = 2
<error descr="Assignment to keyword">False</error> = 1
<error descr="Assignment to keyword">None</error> = 0
<error descr="Assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="assignment to keyword">True</error> = 2
<error descr="assignment to keyword">False</error> = 1
<error descr="assignment to keyword">None</error> = 0
<error descr="assignment to keyword">__debug__</error> = -1
() = []
[] = []
Fix test broken by change in message capitalization
GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="Assignment to keyword">True</error> = 2
<error descr="Assignment to keyword">False</error> = 1
<error descr="Assignment to keyword">None</error> = 0
<error descr="Assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
<commit_before><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="assignment to keyword">True</error> = 2
<error descr="assignment to keyword">False</error> = 1
<error descr="assignment to keyword">None</error> = 0
<error descr="assignment to keyword">__debug__</error> = -1
() = []
[] = []
<commit_msg>Fix test broken by change in message capitalization
GitOrigin-RevId: 48f81146490e925712005aa91064e56670abfbda<commit_after><error descr="starred assignment target must be in a list or tuple">*hi</error> = [1, 2]
*a, = range(5)
for a, *b in [(1, 2, 3), (4, 5, 6, 7)]:
print(b)
a, b, c = seq[0], seq[1:-1], seq[-1]
a, *b, c = seq
[a, *b, c] = seq
<error descr="Assignment to keyword">True</error> = 2
<error descr="Assignment to keyword">False</error> = 1
<error descr="Assignment to keyword">None</error> = 0
<error descr="Assignment to keyword">__debug__</error> = -1
() = []
[] = []
|
48c9b0fc46da538633e7597bb919ac15e4accf7c
|
zeus/db/func.py
|
zeus/db/func.py
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
padding = tuple((len(self.cols) - len(elems)) * (None,))
return [tuple(self._param_regexp.findall(e)) + padding for e in elems]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
num_cols = len(self.cols)
padding = (None,) * num_cols
return [
(tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems
]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
Correct padding on array aggregations
|
fix: Correct padding on array aggregations
This was incorrectly building padding based on a single row.
|
Python
|
apache-2.0
|
getsentry/zeus,getsentry/zeus,getsentry/zeus,getsentry/zeus
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
padding = tuple((len(self.cols) - len(elems)) * (None,))
return [tuple(self._param_regexp.findall(e)) + padding for e in elems]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
fix: Correct padding on array aggregations
This was incorrectly building padding based on a single row.
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
num_cols = len(self.cols)
padding = (None,) * num_cols
return [
(tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems
]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
<commit_before>import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
padding = tuple((len(self.cols) - len(elems)) * (None,))
return [tuple(self._param_regexp.findall(e)) + padding for e in elems]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
<commit_msg>fix: Correct padding on array aggregations
This was incorrectly building padding based on a single row.<commit_after>
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
num_cols = len(self.cols)
padding = (None,) * num_cols
return [
(tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems
]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
padding = tuple((len(self.cols) - len(elems)) * (None,))
return [tuple(self._param_regexp.findall(e)) + padding for e in elems]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
fix: Correct padding on array aggregations
This was incorrectly building padding based on a single row.import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
num_cols = len(self.cols)
padding = (None,) * num_cols
return [
(tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems
]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
<commit_before>import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) fo t he returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
padding = tuple((len(self.cols) - len(elems)) * (None,))
return [tuple(self._param_regexp.findall(e)) + padding for e in elems]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
<commit_msg>fix: Correct padding on array aggregations
This was incorrectly building padding based on a single row.<commit_after>import re
from sqlalchemy.sql import func
from sqlalchemy.types import String, TypeDecorator
# https://bitbucket.org/zzzeek/sqlalchemy/issues/3729/using-array_agg-around-row-function-does
class ArrayOfRecord(TypeDecorator):
_array_regexp = re.compile(r"^\{(\".+?\")*\}$")
_chunk_regexp = re.compile(r'"(.*?)",?')
_param_regexp = re.compile(r"[^\(\),]+")
impl = String
def __init__(self, cols):
self.cols = cols
super().__init__()
def process_result_value(self, value, dialect):
# XXX(dcramer): if the trailing value(s?) of the returning array are NULL, postgres seems to
# not return them, and thus our output array does not match the same length as our column
# selection array
#
# For example if the input is:
# ARRAY_AGG_RESULT(col1, col2)
# And the value of col2 is NULL
# The resulting return value from this query will be:
# ({col1_value},)
elems = self._array_regexp.match(value).group(1)
elems = [e for e in self._chunk_regexp.split(elems) if e]
num_cols = len(self.cols)
padding = (None,) * num_cols
return [
(tuple(self._param_regexp.findall(e)) + padding)[:num_cols] for e in elems
]
def array_agg_row(*arg):
return func.array_agg(func.row(*arg), type_=ArrayOfRecord(arg))
|
83042027fe74ffe200d0bdaa79b0529af54ae6dc
|
addons/website/__openerp__.py
|
addons/website/__openerp__.py
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['base_setup', 'web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
Revert "[FIX] website: add missing module dependency `base_setup`"
|
Revert "[FIX] website: add missing module dependency `base_setup`"
This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1.
The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.
|
Python
|
agpl-3.0
|
ygol/odoo,dfang/odoo,hip-odoo/odoo,hip-odoo/odoo,ygol/odoo,ygol/odoo,dfang/odoo,dfang/odoo,ygol/odoo,hip-odoo/odoo,hip-odoo/odoo,ygol/odoo,ygol/odoo,ygol/odoo,dfang/odoo,hip-odoo/odoo,hip-odoo/odoo,dfang/odoo,dfang/odoo
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['base_setup', 'web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
Revert "[FIX] website: add missing module dependency `base_setup`"
This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1.
The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
<commit_before># -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['base_setup', 'web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
<commit_msg>Revert "[FIX] website: add missing module dependency `base_setup`"
This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1.
The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.<commit_after>
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['base_setup', 'web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
Revert "[FIX] website: add missing module dependency `base_setup`"
This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1.
The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.# -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
<commit_before># -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['base_setup', 'web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
<commit_msg>Revert "[FIX] website: add missing module dependency `base_setup`"
This reverts commit d269eb0eb62d88e02c4fa33b84178d0e73d82ef1.
The issue has been fixed in 61f2c90d507645492e1904c1005e8da6253788ea.<commit_after># -*- encoding: utf-8 -*-
# Part of Odoo. See LICENSE file for full copyright and licensing details.
{
'name': 'Website Builder',
'category': 'Website',
'sequence': 50,
'summary': 'Build Your Enterprise Website',
'website': 'https://www.odoo.com/page/website-builder',
'version': '1.0',
'description': """
Odoo Website CMS
===================
""",
'depends': ['web', 'web_editor', 'web_planner'],
'installable': True,
'data': [
'data/website_data.xml',
'data/web_planner_data.xml',
'security/ir.model.access.csv',
'security/website_security.xml',
'views/website_templates.xml',
'views/website_navbar_templates.xml',
'views/snippets.xml',
'views/website_views.xml',
'views/res_config_views.xml',
'views/ir_actions_views.xml',
'wizard/base_language_install_views.xml',
],
'demo': [
'data/website_demo.xml',
],
'qweb': ['static/src/xml/website.backend.xml'],
'application': True,
}
|
c138adaf69f5029209f03cafe72f1082cdb78f30
|
ppp_nlp_ml_standalone/requesthandler.py
|
ppp_nlp_ml_standalone/requesthandler.py
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
a, b, c = extract_triplet.extract_from_sentence(sentence)
if a == '?':
subject = ppp_datamodel.Missing()
else:
subject = ppp_datamodel.Resource(value=a)
if b == '?':
predicate = ppp_datamodel.Missing()
else:
predicate = ppp_datamodel.Resource(value=b)
if c == '?':
object = ppp_datamodel.Missing()
else:
object = ppp_datamodel.Resource(value=b)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence, Missing, Resource
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
def missing_or_resource(x):
return Missing() if x == '?' else Resource(value=x)
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
triple = extract_triplet.extract_from_sentence(sentence)
(subject, predicate, object) = map(missing_or_resource, triple)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
Make RequestHandler's code less redundant.
|
Make RequestHandler's code less redundant.
|
Python
|
mit
|
ProjetPP/PPP-QuestionParsing-ML-Standalone,ProjetPP/PPP-QuestionParsing-ML-Standalone
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
a, b, c = extract_triplet.extract_from_sentence(sentence)
if a == '?':
subject = ppp_datamodel.Missing()
else:
subject = ppp_datamodel.Resource(value=a)
if b == '?':
predicate = ppp_datamodel.Missing()
else:
predicate = ppp_datamodel.Resource(value=b)
if c == '?':
object = ppp_datamodel.Missing()
else:
object = ppp_datamodel.Resource(value=b)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]Make RequestHandler's code less redundant.
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence, Missing, Resource
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
def missing_or_resource(x):
return Missing() if x == '?' else Resource(value=x)
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
triple = extract_triplet.extract_from_sentence(sentence)
(subject, predicate, object) = map(missing_or_resource, triple)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
<commit_before>"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
a, b, c = extract_triplet.extract_from_sentence(sentence)
if a == '?':
subject = ppp_datamodel.Missing()
else:
subject = ppp_datamodel.Resource(value=a)
if b == '?':
predicate = ppp_datamodel.Missing()
else:
predicate = ppp_datamodel.Resource(value=b)
if c == '?':
object = ppp_datamodel.Missing()
else:
object = ppp_datamodel.Resource(value=b)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]<commit_msg>Make RequestHandler's code less redundant.<commit_after>
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence, Missing, Resource
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
def missing_or_resource(x):
return Missing() if x == '?' else Resource(value=x)
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
triple = extract_triplet.extract_from_sentence(sentence)
(subject, predicate, object) = map(missing_or_resource, triple)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
a, b, c = extract_triplet.extract_from_sentence(sentence)
if a == '?':
subject = ppp_datamodel.Missing()
else:
subject = ppp_datamodel.Resource(value=a)
if b == '?':
predicate = ppp_datamodel.Missing()
else:
predicate = ppp_datamodel.Resource(value=b)
if c == '?':
object = ppp_datamodel.Missing()
else:
object = ppp_datamodel.Resource(value=b)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]Make RequestHandler's code less redundant."""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence, Missing, Resource
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
def missing_or_resource(x):
return Missing() if x == '?' else Resource(value=x)
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
triple = extract_triplet.extract_from_sentence(sentence)
(subject, predicate, object) = map(missing_or_resource, triple)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
<commit_before>"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
a, b, c = extract_triplet.extract_from_sentence(sentence)
if a == '?':
subject = ppp_datamodel.Missing()
else:
subject = ppp_datamodel.Resource(value=a)
if b == '?':
predicate = ppp_datamodel.Missing()
else:
predicate = ppp_datamodel.Resource(value=b)
if c == '?':
object = ppp_datamodel.Missing()
else:
object = ppp_datamodel.Resource(value=b)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]<commit_msg>Make RequestHandler's code less redundant.<commit_after>"""Request handler of the module."""
import ppp_datamodel
from ppp_datamodel import Sentence, Missing, Resource
from ppp_datamodel.communication import TraceItem, Response
from ppp_nlp_ml_standalone import ExtractTriplet
def missing_or_resource(x):
return Missing() if x == '?' else Resource(value=x)
class RequestHandler:
def __init__(self, request):
self.request = request
def answer(self):
if not isinstance(self.request.tree, Sentence):
return []
sentence = self.request.tree.value
extract_triplet = ExtractTriplet.ExtractTriplet()
triple = extract_triplet.extract_from_sentence(sentence)
(subject, predicate, object) = map(missing_or_resource, triple)
triple = ppp_datamodel.Triple(subject=subject,
predicate=predicate,
object=object)
meas = {'accuracy': 0.5, 'relevance': 0.5}
trace = self.request.trace + [TraceItem('NLP-ML-standalone', triple, meas)]
response = Response('en', triple, meas, trace)
print(repr(response))
return [response]
|
b4473d45ba5925551334762bc02708fcb373c957
|
config.py
|
config.py
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False
}
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False,
'page_size': 100
}
|
Add page size for sql query
|
Add page size for sql query
|
Python
|
mit
|
goodwinxp/ATFGenerator,goodwinxp/ATFGenerator,goodwinxp/ATFGenerator
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False
}
Add page size for sql query
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False,
'page_size': 100
}
|
<commit_before>CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False
}
<commit_msg>Add page size for sql query<commit_after>
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False,
'page_size': 100
}
|
CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False
}
Add page size for sql queryCONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False,
'page_size': 100
}
|
<commit_before>CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False
}
<commit_msg>Add page size for sql query<commit_after>CONFIG = {
'database': './ida_info.sqlite3',
'out_dir': './code_gen/',
'verbose': False,
'page_size': 100
}
|
840c764dbe8da07546381b51996f64f7680327bb
|
sciunit/unit_test/doc_tests.py
|
sciunit/unit_test/doc_tests.py
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
def test_chapter6(self):
self.do_notebook("chapter6")
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
#def test_chapter6(self):
# self.do_notebook("chapter6")
|
Drop test ch. 6 since sympy is not in travis-ci
|
Drop test ch. 6 since sympy is not in travis-ci
|
Python
|
mit
|
scidash/sciunit,scidash/sciunit
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
def test_chapter6(self):
self.do_notebook("chapter6")
Drop test ch. 6 since sympy is not in travis-ci
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
#def test_chapter6(self):
# self.do_notebook("chapter6")
|
<commit_before>"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
def test_chapter6(self):
self.do_notebook("chapter6")
<commit_msg>Drop test ch. 6 since sympy is not in travis-ci<commit_after>
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
#def test_chapter6(self):
# self.do_notebook("chapter6")
|
"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
def test_chapter6(self):
self.do_notebook("chapter6")
Drop test ch. 6 since sympy is not in travis-ci"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
#def test_chapter6(self):
# self.do_notebook("chapter6")
|
<commit_before>"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
def test_chapter6(self):
self.do_notebook("chapter6")
<commit_msg>Drop test ch. 6 since sympy is not in travis-ci<commit_after>"""Unit tests for documentation"""
import unittest
from sciunit.utils import NotebookTools
class DocumentationTestCase(NotebookTools, unittest.TestCase):
"""Unit tests for documentation notebooks"""
path = "../../docs"
def test_chapter1(self):
self.do_notebook("chapter1")
def test_chapter2(self):
self.do_notebook("chapter2")
def test_chapter3(self):
self.do_notebook("chapter3")
def test_chapter4(self):
self.do_notebook("chapter4")
def test_chapter5(self):
self.do_notebook("chapter5")
#def test_chapter6(self):
# self.do_notebook("chapter6")
|
cc19d0af1c22c9677960f406ced425aa48da54c1
|
src/sentry/migrations/0063_remove_bad_groupedmessage_index.py
|
src/sentry/migrations/0063_remove_bad_groupedmessage_index.py
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except Exception:
db.rollback_transaction()
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
Revert "Dont error if 0063 index was already cleaned up"
|
Revert "Dont error if 0063 index was already cleaned up"
This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.
|
Python
|
bsd-3-clause
|
felixbuenemann/sentry,JackDanger/sentry,zenefits/sentry,korealerts1/sentry,fuziontech/sentry,daevaorn/sentry,argonemyth/sentry,beeftornado/sentry,vperron/sentry,mvaled/sentry,rdio/sentry,gg7/sentry,hongliang5623/sentry,felixbuenemann/sentry,jokey2k/sentry,pauloschilling/sentry,beni55/sentry,rdio/sentry,BayanGroup/sentry,ngonzalvez/sentry,gencer/sentry,mitsuhiko/sentry,Natim/sentry,gg7/sentry,boneyao/sentry,argonemyth/sentry,NickPresta/sentry,llonchj/sentry,SilentCircle/sentry,looker/sentry,Kryz/sentry,ngonzalvez/sentry,daevaorn/sentry,jokey2k/sentry,alexm92/sentry,BuildingLink/sentry,JamesMura/sentry,pauloschilling/sentry,drcapulet/sentry,beni55/sentry,camilonova/sentry,looker/sentry,vperron/sentry,JamesMura/sentry,mvaled/sentry,songyi199111/sentry,fotinakis/sentry,1tush/sentry,gencer/sentry,jean/sentry,wujuguang/sentry,ifduyue/sentry,JTCunning/sentry,fuziontech/sentry,vperron/sentry,mvaled/sentry,fuziontech/sentry,beeftornado/sentry,Kryz/sentry,camilonova/sentry,ewdurbin/sentry,kevinastone/sentry,llonchj/sentry,gencer/sentry,nicholasserra/sentry,jean/sentry,TedaLIEz/sentry,drcapulet/sentry,mvaled/sentry,looker/sentry,imankulov/sentry,gg7/sentry,NickPresta/sentry,songyi199111/sentry,mvaled/sentry,SilentCircle/sentry,JamesMura/sentry,BuildingLink/sentry,kevinlondon/sentry,Natim/sentry,hongliang5623/sentry,wujuguang/sentry,felixbuenemann/sentry,kevinastone/sentry,gencer/sentry,daevaorn/sentry,JamesMura/sentry,korealerts1/sentry,korealerts1/sentry,alexm92/sentry,BuildingLink/sentry,ifduyue/sentry,looker/sentry,beeftornado/sentry,JackDanger/sentry,BayanGroup/sentry,gencer/sentry,fotinakis/sentry,songyi199111/sentry,beni55/sentry,nicholasserra/sentry,Natim/sentry,zenefits/sentry,hongliang5623/sentry,1tush/sentry,argonemyth/sentry,llonchj/sentry,alexm92/sentry,zenefits/sentry,JackDanger/sentry,ifduyue/sentry,1tush/sentry,boneyao/sentry,JamesMura/sentry,SilentCircle/sentry,pauloschilling/sentry,ewdurbin/sentry,zenefits/sentry,mvaled/sentry,imankulov/sentry,mitsuhiko/sentry,Kryz/sentry,looker/sentry,kevinastone/sentry,kevinlondon/sentry,zenefits/sentry,ifduyue/sentry,NickPresta/sentry,wong2/sentry,ngonzalvez/sentry,rdio/sentry,fotinakis/sentry,jokey2k/sentry,SilentCircle/sentry,TedaLIEz/sentry,fotinakis/sentry,nicholasserra/sentry,daevaorn/sentry,TedaLIEz/sentry,camilonova/sentry,wong2/sentry,BuildingLink/sentry,jean/sentry,JTCunning/sentry,imankulov/sentry,jean/sentry,ifduyue/sentry,JTCunning/sentry,NickPresta/sentry,kevinlondon/sentry,drcapulet/sentry,rdio/sentry,wujuguang/sentry,jean/sentry,BayanGroup/sentry,boneyao/sentry,ewdurbin/sentry,wong2/sentry,BuildingLink/sentry
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except Exception:
db.rollback_transaction()
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
Revert "Dont error if 0063 index was already cleaned up"
This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
<commit_before># encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except Exception:
db.rollback_transaction()
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
<commit_msg>Revert "Dont error if 0063 index was already cleaned up"
This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.<commit_after>
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except Exception:
db.rollback_transaction()
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
Revert "Dont error if 0063 index was already cleaned up"
This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.# encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
<commit_before># encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
try:
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
except Exception:
db.rollback_transaction()
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
<commit_msg>Revert "Dont error if 0063 index was already cleaned up"
This reverts commit b3a51fa482fc949de75d962ddd9fe3464fa70e58.<commit_after># encoding: utf-8
import datetime
from south.db import db
from south.v2 import SchemaMigration
from django.db import models
class Migration(SchemaMigration):
def forwards(self, orm):
# Removing unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.delete_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
def backwards(self, orm):
# Adding unique constraint on 'GroupedMessage', fields ['logger', 'view', 'checksum']
db.create_unique('sentry_groupedmessage', ['logger', 'view', 'checksum'])
|
da0d5d5058e57b3f2dd2b16822e0ab47ced7353a
|
AddressServer.py
|
AddressServer.py
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
SocketServer.TCPServer.allow_reuse_address = True
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
Allow re-use of a port a la SO_REUSEADDRESS
|
Allow re-use of a port a la SO_REUSEADDRESS
|
Python
|
apache-2.0
|
mattrichards/TCPAddressServer
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
Allow re-use of a port a la SO_REUSEADDRESS
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
SocketServer.TCPServer.allow_reuse_address = True
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
<commit_msg>Allow re-use of a port a la SO_REUSEADDRESS<commit_after>
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
SocketServer.TCPServer.allow_reuse_address = True
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
Allow re-use of a port a la SO_REUSEADDRESS#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
SocketServer.TCPServer.allow_reuse_address = True
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
<commit_before>#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
<commit_msg>Allow re-use of a port a la SO_REUSEADDRESS<commit_after>#!/usr/bin/python
"""
Simple TCP server that sends back the IP address and port from which it
sees you connect. Useful when used with `telnet` for network debugging.
Currently only IPv4.
"""
import SocketServer
VERBOSE = False
class TCPHandler(SocketServer.BaseRequestHandler):
"""
Handles each request be sending the client's IP address and port, then
closing the connection.
"""
def handle(self):
# self.request is the TCP socket connected to the client
address = self.client_address[0] + ":" + str(self.client_address[1])
if VERBOSE:
print "Request from: " + address
self.request.sendall(address + "\n")
def main():
import optparse
global VERBOSE
parser = optparse.OptionParser()
parser.add_option("-p", "--port", type="int", default=9999,
help="port to listen on")
parser.add_option("-v", "--verbose", action="store_true")
options = parser.parse_args()[0]
port = options.port
VERBOSE = options.verbose
SocketServer.TCPServer.allow_reuse_address = True
server = SocketServer.TCPServer(("0.0.0.0", port), TCPHandler)
print "Listening on port %d" % port
# Activate the server; this will keep running until you
# interrupt the program with Ctrl-C
server.serve_forever()
if __name__ == "__main__":
main()
|
1daeedde2cd8597e047b6a6d7fc737f103fa4ac8
|
example/handler/my_handler.py
|
example/handler/my_handler.py
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
print("Hello there, {}!".format(who))
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload, with_member_info
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
"""A custom user event."""
print("Hello there, {}!".format(who))
@with_payload
def supervisor(self, **kwargs):
"""This event gets fired for change of state in supervisor."""
print("Got a supervisor event with payload:")
print(kwargs)
@with_member_info
def member_join(self, members):
"""This event gets fired on members joining.
``members`` is the dict with new members'
"""
print("Hello to new members:")
print(members)
|
Add example of supervisor event and member join
|
Add example of supervisor event and member join
|
Python
|
mit
|
waltermoreira/serfnode,waltermoreira/serfnode,waltermoreira/serfnode
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
print("Hello there, {}!".format(who))
Add example of supervisor event and member join
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload, with_member_info
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
"""A custom user event."""
print("Hello there, {}!".format(who))
@with_payload
def supervisor(self, **kwargs):
"""This event gets fired for change of state in supervisor."""
print("Got a supervisor event with payload:")
print(kwargs)
@with_member_info
def member_join(self, members):
"""This event gets fired on members joining.
``members`` is the dict with new members'
"""
print("Hello to new members:")
print(members)
|
<commit_before>from base_handler import BaseHandler
from utils import truncated_stdout, with_payload
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
print("Hello there, {}!".format(who))
<commit_msg>Add example of supervisor event and member join<commit_after>
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload, with_member_info
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
"""A custom user event."""
print("Hello there, {}!".format(who))
@with_payload
def supervisor(self, **kwargs):
"""This event gets fired for change of state in supervisor."""
print("Got a supervisor event with payload:")
print(kwargs)
@with_member_info
def member_join(self, members):
"""This event gets fired on members joining.
``members`` is the dict with new members'
"""
print("Hello to new members:")
print(members)
|
from base_handler import BaseHandler
from utils import truncated_stdout, with_payload
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
print("Hello there, {}!".format(who))
Add example of supervisor event and member joinfrom base_handler import BaseHandler
from utils import truncated_stdout, with_payload, with_member_info
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
"""A custom user event."""
print("Hello there, {}!".format(who))
@with_payload
def supervisor(self, **kwargs):
"""This event gets fired for change of state in supervisor."""
print("Got a supervisor event with payload:")
print(kwargs)
@with_member_info
def member_join(self, members):
"""This event gets fired on members joining.
``members`` is the dict with new members'
"""
print("Hello to new members:")
print(members)
|
<commit_before>from base_handler import BaseHandler
from utils import truncated_stdout, with_payload
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
print("Hello there, {}!".format(who))
<commit_msg>Add example of supervisor event and member join<commit_after>from base_handler import BaseHandler
from utils import truncated_stdout, with_payload, with_member_info
class MyHandler(BaseHandler):
@truncated_stdout
@with_payload
def hello(self, who=None):
"""A custom user event."""
print("Hello there, {}!".format(who))
@with_payload
def supervisor(self, **kwargs):
"""This event gets fired for change of state in supervisor."""
print("Got a supervisor event with payload:")
print(kwargs)
@with_member_info
def member_join(self, members):
"""This event gets fired on members joining.
``members`` is the dict with new members'
"""
print("Hello to new members:")
print(members)
|
54d4551ce8efb16d4a8d02e38b9f223f8f1cd816
|
ab_game.py
|
ab_game.py
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
CAPTURE_SCORE_BASE = 120 ** 3
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
if depth > 4:
min_priority = 3
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
Disable min_priority filter for now
|
Disable min_priority filter for now
|
Python
|
mit
|
cropleyb/pentai,cropleyb/pentai,cropleyb/pentai
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
CAPTURE_SCORE_BASE = 120 ** 3
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
if depth > 4:
min_priority = 3
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
Disable min_priority filter for now
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
<commit_before>#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
CAPTURE_SCORE_BASE = 120 ** 3
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
if depth > 4:
min_priority = 3
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
<commit_msg>Disable min_priority filter for now<commit_after>
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
CAPTURE_SCORE_BASE = 120 ** 3
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
if depth > 4:
min_priority = 3
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
Disable min_priority filter for now#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
<commit_before>#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
CAPTURE_SCORE_BASE = 120 ** 3
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
if depth > 4:
min_priority = 3
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
<commit_msg>Disable min_priority filter for now<commit_after>#!/usr/bin/python
import board
import pente_exceptions
from ab_state import *
class ABGame():
""" This class acts as a bridge between the AlphaBeta code and my code """
def __init__(self, base_game):
s = self.current_state = ABState()
s.set_state(base_game.current_state)
self.base_game = base_game
def to_move(self, state=None):
if state is None:
state = self.current_state
return state.to_move()
def utility(self, state):
return state.utility()
def successors(self, state, depth):
mn = state.get_move_number()
if mn == 1:
# The first black move is always in the centre
brd_size = self.base_game.get_board().get_size()
centre_pos = (brd_size/2, brd_size/2)
p_i = [centre_pos]
else:
min_priority = 0
pos_iter = state.get_iter(state.to_move())
p_i = pos_iter.get_iter(state.to_move_colour(), min_priority)
tried_count = 0
for pos in p_i:
# create an AB_State for each possible move from state
succ = state.create_state(pos)
yield pos, succ
tried_count += 1
if depth > 3 and tried_count >= 2:
return
def terminal_test(self, state):
return state.terminal()
|
06d9171b2244e4dd9d5e1883101d7ec3e05be4b2
|
bitfield/apps.py
|
bitfield/apps.py
|
from django.apps import AppConfig
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
import django
from django.apps import AppConfig
django.setup()
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
Add django.setup to the AppConfig
|
Add django.setup to the AppConfig
|
Python
|
apache-2.0
|
Elec/django-bitfield,disqus/django-bitfield,joshowen/django-bitfield
|
from django.apps import AppConfig
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
Add django.setup to the AppConfig
|
import django
from django.apps import AppConfig
django.setup()
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
<commit_before>from django.apps import AppConfig
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
<commit_msg>Add django.setup to the AppConfig<commit_after>
|
import django
from django.apps import AppConfig
django.setup()
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
from django.apps import AppConfig
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
Add django.setup to the AppConfigimport django
from django.apps import AppConfig
django.setup()
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
<commit_before>from django.apps import AppConfig
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
<commit_msg>Add django.setup to the AppConfig<commit_after>import django
from django.apps import AppConfig
django.setup()
class BitFieldAppConfig(AppConfig):
name = 'bitfield'
verbose_name = "Bit Field"
|
0db43d894bfb419a7f4b538f755af47fc0b653cb
|
tests/unit/test_sharpspring.py
|
tests/unit/test_sharpspring.py
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
Add mock response to sharpspring test
|
Add mock response to sharpspring test
|
Python
|
apache-2.0
|
Code4SA/pmg-cms-2,Code4SA/pmg-cms-2,Code4SA/pmg-cms-2
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
Add mock response to sharpspring test
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
<commit_before>from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
<commit_msg>Add mock response to sharpspring test<commit_after>
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
Add mock response to sharpspring testfrom unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
<commit_before>from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post")
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
<commit_msg>Add mock response to sharpspring test<commit_after>from unittest.mock import patch
from pmg.sharpspring import Sharpspring
from tests import PMGTestCase
class MockResponse:
def __init__(self, json_data, status_code):
self.json_data = json_data
self.status_code = status_code
def raise_for_status(self):
pass
def json(self):
return self.json_data
def mocked_requests_post_success(*args, **kwargs):
return MockResponse({"result": {"creates": [{"success": True,}]}}, 200)
class TestSharpspring(PMGTestCase):
@patch("pmg.sharpspring.requests.post", side_effect=mocked_requests_post_success)
def test_make_sharpsrping_request(self, post_mock):
sharpspring = Sharpspring()
details = {
"emailAddress": "test@example.com",
"companyName": "Test Company",
}
result = sharpspring.call("createLeads", {"objects": [details]})
post_mock.assert_called()
self.assertTrue(result["result"]["creates"][0])
|
0cb807470ee56207251f36ad78d35c48f6e9361b
|
example_project/urls.py
|
example_project/urls.py
|
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
from django.conf.urls import include, url
from django.contrib import admin
admin.autodiscover() # For Django 1.6
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
|
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
|
Python
|
mit
|
BocuStudio/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,BocuStudio/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,caktus/django-timepiece,arbitrahj/django-timepiece,BocuStudio/django-timepiece
|
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility
|
from django.conf.urls import include, url
from django.contrib import admin
admin.autodiscover() # For Django 1.6
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
<commit_before>from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
<commit_msg>Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility<commit_after>
|
from django.conf.urls import include, url
from django.contrib import admin
admin.autodiscover() # For Django 1.6
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibilityfrom django.conf.urls import include, url
from django.contrib import admin
admin.autodiscover() # For Django 1.6
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
<commit_before>from django.conf.urls import include, url
from django.contrib import admin
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
<commit_msg>Update Python/Django: Restore admin.autodiscover() for Django 1.6 compatibility<commit_after>from django.conf.urls import include, url
from django.contrib import admin
admin.autodiscover() # For Django 1.6
urlpatterns = [
url(r'^admin/', include(admin.site.urls)),
url(r'^selectable/', include('selectable.urls')),
url(r'', include('timepiece.urls')),
# authentication views
url(r'^accounts/login/$', 'django.contrib.auth.views.login',
name='auth_login'),
url(r'^accounts/logout/$', 'django.contrib.auth.views.logout_then_login',
name='auth_logout'),
url(r'^accounts/password-change/$',
'django.contrib.auth.views.password_change',
name='change_password'),
url(r'^accounts/password-change/done/$',
'django.contrib.auth.views.password_change_done'),
url(r'^accounts/password-reset/$',
'django.contrib.auth.views.password_reset',
name='reset_password'),
url(r'^accounts/password-reset/done/$',
'django.contrib.auth.views.password_reset_done'),
url(r'^accounts/reset/(?P<uidb36>[0-9A-Za-z]+)-(?P<token>.+)/$',
'django.contrib.auth.views.password_reset_confirm'),
url(r'^accounts/reset/done/$',
'django.contrib.auth.views.password_reset_complete'),
]
|
23343eb3316a3d304a3b021519b9a470f9c2446b
|
django_bcrypt/models.py
|
django_bcrypt/models.py
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
if raw_password is None:
self.set_unusable_password()
else:
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
Allow users to be created with blank (unusable) passwords.
|
Allow users to be created with blank (unusable) passwords.
|
Python
|
mit
|
dwaiter/django-bcrypt
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
Allow users to be created with blank (unusable) passwords.
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
if raw_password is None:
self.set_unusable_password()
else:
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
<commit_before>import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
<commit_msg>Allow users to be created with blank (unusable) passwords.<commit_after>
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
if raw_password is None:
self.set_unusable_password()
else:
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
Allow users to be created with blank (unusable) passwords.import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
if raw_password is None:
self.set_unusable_password()
else:
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
<commit_before>import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
<commit_msg>Allow users to be created with blank (unusable) passwords.<commit_after>import bcrypt
from django.contrib.auth.models import User
from django.conf import settings
try:
rounds = settings.BCRYPT_ROUNDS
except AttributeError:
rounds = 12
_check_password = User.check_password
def bcrypt_check_password(self, raw_password):
if self.password.startswith('bc$'):
salt_and_hash = self.password[3:]
return bcrypt.hashpw(raw_password, salt_and_hash) == salt_and_hash
return _check_password(self, raw_password)
def bcrypt_set_password(self, raw_password):
if raw_password is None:
self.set_unusable_password()
else:
salt = bcrypt.gensalt(rounds)
self.password = 'bc$' + bcrypt.hashpw(raw_password, salt)
User.check_password = bcrypt_check_password
User.set_password = bcrypt_set_password
|
f5e4a8000e23e279192834d03e4b5b9ecca6b2b0
|
linguist/utils/__init__.py
|
linguist/utils/__init__.py
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
get_real_field_name,
get_fallback_field_name,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
Fix new i18n utils imports.
|
Fix new i18n utils imports.
|
Python
|
mit
|
ulule/django-linguist
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
Fix new i18n utils imports.
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
get_real_field_name,
get_fallback_field_name,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
<commit_before># -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
<commit_msg>Fix new i18n utils imports.<commit_after>
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
get_real_field_name,
get_fallback_field_name,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
Fix new i18n utils imports.# -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
get_real_field_name,
get_fallback_field_name,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
<commit_before># -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
<commit_msg>Fix new i18n utils imports.<commit_after># -*- coding: utf-8 -*-
from .i18n import (get_language_name,
get_language,
get_fallback_language,
get_real_field_name,
get_fallback_field_name,
build_localized_field_name,
build_localized_verbose_name)
from .models import load_class, get_model_string
from .template import select_template_name
from .views import get_language_parameter, get_language_tabs
__all__ = [
'get_language_name',
'get_language',
'get_fallback_language',
'build_localized_field_name',
'build_localized_verbose_name',
'load_class',
'get_model_string',
'select_template_name',
'get_language_parameter',
'get_language_tabs',
'chunks',
]
def chunks(l, n):
"""
Yields successive n-sized chunks from l.
"""
for i in xrange(0, len(l), n):
yield l[i:i + n]
|
986b9227fe66d95a7e42253395c89de5c2385b2d
|
scuole/campuses/management/commands/dedupecampusslugs.py
|
scuole/campuses/management/commands/dedupecampusslugs.py
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
print('DUPLICATE SLUGS', duplicate_slugs)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
# for campus in Campus.objects.filter(slug=slug):
# if campus.city != None:
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# campus.save()
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# print(slugify(campus.city, allow_unicode=True))
# print('SLUG', campus.slug)
# campus.save()
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
campuses_dup_slug = Campus.objects.filter(slug=slug)
# if the district and county are the same, but the city of the campuses are different
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
if campus.city != None:
city_slug = slugify(campus.city, allow_unicode=True)
campus.slug = f"{campus.slug}-{city_slug}"
campus.save()
# if the district, county, and city of the campuses are the same
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
campus.slug = f"{campus.slug}-{campus.tea_id}"
campus.save()
|
Edit dedupe campus slugs code
|
Edit dedupe campus slugs code
|
Python
|
mit
|
texastribune/scuole,texastribune/scuole,texastribune/scuole,texastribune/scuole
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
print('DUPLICATE SLUGS', duplicate_slugs)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
# for campus in Campus.objects.filter(slug=slug):
# if campus.city != None:
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# campus.save()
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# print(slugify(campus.city, allow_unicode=True))
# print('SLUG', campus.slug)
# campus.save()
Edit dedupe campus slugs code
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
campuses_dup_slug = Campus.objects.filter(slug=slug)
# if the district and county are the same, but the city of the campuses are different
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
if campus.city != None:
city_slug = slugify(campus.city, allow_unicode=True)
campus.slug = f"{campus.slug}-{city_slug}"
campus.save()
# if the district, county, and city of the campuses are the same
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
campus.slug = f"{campus.slug}-{campus.tea_id}"
campus.save()
|
<commit_before>from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
print('DUPLICATE SLUGS', duplicate_slugs)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
# for campus in Campus.objects.filter(slug=slug):
# if campus.city != None:
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# campus.save()
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# print(slugify(campus.city, allow_unicode=True))
# print('SLUG', campus.slug)
# campus.save()
<commit_msg>Edit dedupe campus slugs code<commit_after>
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
campuses_dup_slug = Campus.objects.filter(slug=slug)
# if the district and county are the same, but the city of the campuses are different
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
if campus.city != None:
city_slug = slugify(campus.city, allow_unicode=True)
campus.slug = f"{campus.slug}-{city_slug}"
campus.save()
# if the district, county, and city of the campuses are the same
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
campus.slug = f"{campus.slug}-{campus.tea_id}"
campus.save()
|
from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
print('DUPLICATE SLUGS', duplicate_slugs)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
# for campus in Campus.objects.filter(slug=slug):
# if campus.city != None:
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# campus.save()
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# print(slugify(campus.city, allow_unicode=True))
# print('SLUG', campus.slug)
# campus.save()
Edit dedupe campus slugs codefrom django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
campuses_dup_slug = Campus.objects.filter(slug=slug)
# if the district and county are the same, but the city of the campuses are different
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
if campus.city != None:
city_slug = slugify(campus.city, allow_unicode=True)
campus.slug = f"{campus.slug}-{city_slug}"
campus.save()
# if the district, county, and city of the campuses are the same
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
campus.slug = f"{campus.slug}-{campus.tea_id}"
campus.save()
|
<commit_before>from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
print('DUPLICATE SLUGS', duplicate_slugs)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
# for campus in Campus.objects.filter(slug=slug):
# if campus.city != None:
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# campus.save()
# city_slug = slugify(campus.city, allow_unicode=True)
# campus.slug = f"{campus.slug}-{city_slug}"
# print(slugify(campus.city, allow_unicode=True))
# print('SLUG', campus.slug)
# campus.save()
<commit_msg>Edit dedupe campus slugs code<commit_after>from django.core.management.base import BaseCommand
from django.db.models import Count
from django.utils.text import slugify
from scuole.campuses.models import Campus
class Command(BaseCommand):
help = "Dedupe Campus slugs by adding the county name to the end."
def handle(self, *args, **options):
duplicate_slugs = (
Campus.objects.values("slug")
.annotate(total=Count("slug"))
.filter(total__gt=1)
)
# loop through all duplicate slugs
for duplicate in duplicate_slugs:
slug = duplicate['slug']
campuses_dup_slug = Campus.objects.filter(slug=slug)
# if the district and county are the same, but the city of the campuses are different
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
if campus.city != None:
city_slug = slugify(campus.city, allow_unicode=True)
campus.slug = f"{campus.slug}-{city_slug}"
campus.save()
# if the district, county, and city of the campuses are the same
if all(obj.district == campuses_dup_slug[0].district for obj in campuses_dup_slug) and all(obj.county == campuses_dup_slug[0].county for obj in campuses_dup_slug) and all(obj.city == campuses_dup_slug[0].city for obj in campuses_dup_slug):
for campus in campuses_dup_slug:
campus.slug = f"{campus.slug}-{campus.tea_id}"
campus.save()
|
de0bb4886b9a6ecd2fb4e5c4272167911141c71c
|
apic_ml2/neutron/plugins/ml2/drivers/cisco/apic/nova_client.py
|
apic_ml2/neutron/plugins/ml2/drivers/cisco/apic/nova_client.py
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
class NovaClient(object):
def __init__(self):
self.nclient = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.nclient.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
client = None
def _get_client():
global client
if client is None:
client = n_nova.Notifier().nclient
return client
class NovaClient(object):
def __init__(self):
self.client = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
Load Nova Client only once to avoid reconnecting
|
Load Nova Client only once to avoid reconnecting
|
Python
|
apache-2.0
|
noironetworks/apic-ml2-driver,noironetworks/apic-ml2-driver
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
class NovaClient(object):
def __init__(self):
self.nclient = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.nclient.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
Load Nova Client only once to avoid reconnecting
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
client = None
def _get_client():
global client
if client is None:
client = n_nova.Notifier().nclient
return client
class NovaClient(object):
def __init__(self):
self.client = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
class NovaClient(object):
def __init__(self):
self.nclient = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.nclient.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
<commit_msg>Load Nova Client only once to avoid reconnecting<commit_after>
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
client = None
def _get_client():
global client
if client is None:
client = n_nova.Notifier().nclient
return client
class NovaClient(object):
def __init__(self):
self.client = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
class NovaClient(object):
def __init__(self):
self.nclient = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.nclient.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
Load Nova Client only once to avoid reconnecting# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
client = None
def _get_client():
global client
if client is None:
client = n_nova.Notifier().nclient
return client
class NovaClient(object):
def __init__(self):
self.client = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
class NovaClient(object):
def __init__(self):
self.nclient = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.nclient.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
<commit_msg>Load Nova Client only once to avoid reconnecting<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from neutron._i18n import _LW
from neutron.notifiers import nova as n_nova
from novaclient import exceptions as nova_exceptions
from oslo_log import log as logging
LOG = logging.getLogger(__name__)
client = None
def _get_client():
global client
if client is None:
client = n_nova.Notifier().nclient
return client
class NovaClient(object):
def __init__(self):
self.client = n_nova.Notifier().nclient
def get_server(self, server_id):
try:
return self.client.servers.get(server_id)
except nova_exceptions.NotFound:
LOG.warning(_LW("Nova returned NotFound for server: %s"),
server_id)
except Exception as e:
LOG.exception(e)
|
79c5a3b12fbe0ccde4bf8ec8694d42696241621d
|
products/bika/browser/clientfolder.py
|
products/bika/browser/clientfolder.py
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'},
'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'},
'getPhone': {'title': 'Phone', 'field':'getPhone'},
'getFax': {'title': 'Fax', 'field':'getFax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name'},
'getEmailAddress': {'title': 'Email Address'},
'getPhone': {'title': 'Phone'},
'getFax': {'title': 'Fax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
Remove 'field' and 'icon' from column list
|
Remove 'field' and 'icon' from column list
|
Python
|
agpl-3.0
|
veroc/Bika-LIMS,veroc/Bika-LIMS,veroc/Bika-LIMS,DeBortoliWines/Bika-LIMS,labsanmartin/Bika-LIMS,rockfruit/bika.lims,labsanmartin/Bika-LIMS,DeBortoliWines/Bika-LIMS,labsanmartin/Bika-LIMS,anneline/Bika-LIMS,anneline/Bika-LIMS,DeBortoliWines/Bika-LIMS,anneline/Bika-LIMS,rockfruit/bika.lims
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'},
'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'},
'getPhone': {'title': 'Phone', 'field':'getPhone'},
'getFax': {'title': 'Fax', 'field':'getFax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
Remove 'field' and 'icon' from column list
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name'},
'getEmailAddress': {'title': 'Email Address'},
'getPhone': {'title': 'Phone'},
'getFax': {'title': 'Fax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
<commit_before>from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'},
'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'},
'getPhone': {'title': 'Phone', 'field':'getPhone'},
'getFax': {'title': 'Fax', 'field':'getFax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
<commit_msg>Remove 'field' and 'icon' from column list<commit_after>
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name'},
'getEmailAddress': {'title': 'Email Address'},
'getPhone': {'title': 'Phone'},
'getFax': {'title': 'Fax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'},
'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'},
'getPhone': {'title': 'Phone', 'field':'getPhone'},
'getFax': {'title': 'Fax', 'field':'getFax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
Remove 'field' and 'icon' from column listfrom Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name'},
'getEmailAddress': {'title': 'Email Address'},
'getPhone': {'title': 'Phone'},
'getFax': {'title': 'Fax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
<commit_before>from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name', 'field': 'title_or_id', 'icon': 'client.png'},
'getEmailAddress': {'title': 'Email Address', 'field':'getEmailAddress'},
'getPhone': {'title': 'Phone', 'field':'getPhone'},
'getFax': {'title': 'Fax', 'field':'getFax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
<commit_msg>Remove 'field' and 'icon' from column list<commit_after>from Products.CMFCore.utils import getToolByName
from Products.bika import logger
from Products.bika.browser.bika_folder_contents import BikaFolderContentsView
from plone.app.content.browser.interfaces import IFolderContentsView
from zope.interface import implements
class ClientFolderContentsView(BikaFolderContentsView):
implements(IFolderContentsView)
contentFilter = {'portal_type': 'Client'}
content_add_buttons = ['Client', ]
batch = True
b_size = 100
show_editable_border = False
columns = {
'title_or_id': {'title': 'Name'},
'getEmailAddress': {'title': 'Email Address'},
'getPhone': {'title': 'Phone'},
'getFax': {'title': 'Fax'},
}
wflist_states = [
{'title': 'All', 'id':'all',
'columns':['title_or_id',
'getEmailAddress',
'getPhone',
'getFax', ],
'buttons':[BikaFolderContentsView.default_buttons['delete']]},
]
def folderitems(self):
items = BikaFolderContentsView.folderitems(self)
for x in range(len(items)):
items[x]['links'] = {'title_or_id': items[x]['url']}
return items
def __call__(self):
return self.template()
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.